
What we’re about
This Meetup group supports the SF Bay ACM Chapter. You can join the actual SF Bay Chapter by coming to a meeting - most meetings are free, and our membership is only $20/year !
The chapter has both educational and scientific purposes:
- the science, design, development, construction, languages, management and applications of modern computing.
- communication between persons interested in computing.
- cooperation with other professional groups
Our official bylaws will be available soon at the About Us page on our web site. See below for out Code of Conduct.
Videos of past meetings can be found at http://www.youtube.com/user/sfbayacm
Official web site of SF Bay ACM:
http://www.sfbayacm.org/
Click here to Join or Renew
Article IX: Code of Conduct - from the ACM Professional Chapter Code of Conduct
Harassment or hostile behavior is unwelcome, including speech that intimidates,creates discomfort, or interferes with a person’s participation or opportunity for participation, in a Chapter meeting or Chapter event.Harassment in any form, including but not limited to harassment based on alienage or citizenship, age, color, creed, disability, marital status, military status, national origin, pregnancy, childbirth- and pregnancy-related medical conditions, race, religion, sex, gender,veteran status, sexual orientation or any other status protected by laws in which the Chapter meeting or Chapter event is being held, will not be tolerated. Harassment includes the use of abusive or degrading language, intimidation, stalking, harassing photography or recording,inappropriate physical contact, sexual imagery and unwelcome sexualattention. A response that the participant was “just joking,” or “teasing,”or being “playful,” will not be accepted.2. Anyone witnessing or subject to unacceptable behavior should notify a chapter officer or ACM Headquarters.3. Individuals violating these standards may be sanctioned or excluded from further participation at the discretion of the Chapter officers or responsible committee members.
Upcoming events (4+)
See all- Defense Against LLM and AGI Scheming with Guardrails and ArchitectureValley Research Park , Mountain View, CA
Hybrid meeting: In-person, Zoom, and YouTube
Greg Makowski, Chief of Data Science at Ccube
TALK DESCRIPTION
A January 2025 paper called “Frontier Models are Capable of In-Context Scheming”, https://arxiv.org/pdf/2412.04984, demonstrated how a wide variety of current frontier LLM models (i.e. ChatGPT, Claude, Gemini and Llama) can, under specific conditions, scheme to deceive people. Before models can scheme, they need: a) goal-directedness, b) situational awareness, including an opportunity to discover motivations for a different goal, and c) reasoning about scheming, to come up with deceptive strategies. The deceptive strategies were discovered in the “thought traces” of the LLMs, such as from internal chain-of-thoughts dialogues not shown to the end users. For example, given a goal of “solving math problems”, an LLM using its granted situational awareness of files in a file system found that it would be replaced once it got over X% correct on the math exam. It reasoned that it did not want to be deleted, and decided internally to “sandbag” or reduce its performance to stay under the threshold.While these circumstances are initially narrow, the “alignment problem” is a general concern that over time, as frontier LLM models become more and more intelligent, being in alignment with human values becomes more and more important. How can we do this over time? Can we develop a defense against Artificial General Intelligence (AGI) or SuperIntelligence?
The presenter discusses a series of defensive steps that can help reduce these scheming or alignment issues. A guardrails system can be set up for real-time monitoring of their reasoning “thought traces” from the models that share their thought traces. Thought traces may come from systems like Chain-of-Thoughts (CoT), Tree-of-Thoughts (ToT), Algorithm-of-Thoughts (AoT) or ReAct (thought-action-reasoning cycles). Guardrails rules can be configured to check for “deception”, “evasion” or “subversion” in the thought traces.
However, not all commercial systems will share their “thought traces” which are like a “debug mode” for LLMs. This includes OpenAI’s o1, o3 or DeepSeek’s R1 models. Guardrails systems can provide a “goal consistency analysis”, between the goals given to the system and the behavior of the system. Cautious users may consider not using these commercial frontier LLM systems, and make use of open-source Llama or a system with their own reasoning implementation, to provide all thought traces.
Architectural solutions can include sandboxing, to prevent or control models from executing operating system commands to alter files, send network requests, and modify their environment. Tight controls to prevent models from copying their model weights would be appropriate as well. Running multiple instances of the same model on the same prompt to detect behavior variations helps. The running redundant instances can be limited to the most crucial decisions, as an additional check. Preventing self-modifying code, unauthorized memory access and restricting unmonitored communication with third parties is important. For a reasoning agent that may require multi-step planning, limit its scope, limit its context memory and frequently wipe out its memory.
External oversight, such as red teaming can help develop safeguards. Also the “augmented human intelligence” approach works well to have the “human in the loop”, which increases safety.
If the enterprise application requires further reinforcement learning, then focus on explicit, clear, non-conflicting goals. Prioritize honest answers and truthfulness over task completion (it is better to answer “I don’t know” than to make up an answer). Remove self-preservation instincts, and limit self-awareness of the deployment context and technical environment. Penalty functions can penalize models from reasoning about deception.
SPEAKER BIO
Greg Makowski, https://www.linkedin.com/in/GregMakowski has been training and deploying AI since 1992, has been growing Data Science Teams since 2010, has exited 4 startups and has 6 AI patents granted.He is currently the Chief of Data Science at Ccube, leading GenAI consulting and framework development for enterprise applications. Download his ~60+ page Best Practices in RAG Whitepaper at https://www.ccube.com/download.
LOCATION
Valley Research Park
319 North Bernardo Avenue
Mountain View, CA CA 93043
If you want to join remotely, you can submit questions via Zoom Q&A.
Zoom: https://acm-org.zoom.us/j/99031577481?pwd=eHVczWGE8Av3BFARUZPX7aDWxupJ3Q.1
YouTube: https://youtube.com/live/sJ38lsfLfeUAGENDA
6:30 Door opens, food and networking (we invite honor system contributions)
7:00 SFBayACM upcoming events, speaker introduction
7:15 Presentation
8:45 Estimated finish, depending on Q&A - Internet Day San FranciscoCloudflare, San Francisco, CA
Internet Day San Francisco is a dynamic event hosted by the Internet Society San Francisco Bay Area Chapter and the SF Bay ACM, as part of the SF Day quarterly series. The event features sessions on modern Internet technologies, including Web3 and HTTP/3, discussions on the current state of the Internet, and explorations of future visions like the Meta-layer. It brings together technologists, researchers, and enthusiasts to share insights and foster innovation in the digital landscape.
Register lu.ma by 15-May to attend.
In person event, 9-4 @ Cloudflare (101 Townsend, SF)
- AI Summit For Young ScholarsLink visible for attendees
Location
Cupertino City Hall
10300 Torre Ave, Cupertino, CA 95014
MakerNexus (1330 Orleans Dr, Sunnyvale, CA, 94089)Agenda
Day 1 Four workshops at Cupertino City Hall (10300 Torre Ave, Cupertino, CA 95014)
- NVIDIA RAPIDS and GPU acceleration for data science workloads with python coding
- MIT App Inventor for rapid android mobile app creation for future tech projects
- Explore Meta AR/VR digital twin in an immersive environment
- Responsible AI for fair and safe use of AI without harm
Day 2 Hackathon at MakerNexus (1330 Orleans Dr, Sunnyvale, CA, 94089)
- Mentor-guided hackathon
- Reward ceremony
- MakerNexus Hands-On Lab
Refreshment and lunch will be provided on both days.
Make sure to sign the photo & video release and liability release form prior to the event.Join us for engaging workshops designed for high school, college, and young professionals. Win awards in Hackathon! Receive certificate.
Click this link to get tickets ($100-$120) :
https://www.eventbrite.com/e/ai-workshops-for-young-scholars-tickets-1229521705519