[devnexus 2026] Hacking AI – How to Survive the AI Uprising

Speaker: Gant Laborde @GantLaborde

See the DevNexus live blog table of contents for more posts


General

  • Can’t blindly trust AI
  • People are trying to put AI in every place possible without thinking through implications

Traditional Hacking

  • Confuse
  • Elevate privileges
  • Destroy

History

  • Captain Crunch whistle – blow into phone and frequency could make free calls long distance
  • Neural Tank Legend – 100% accurate if only ask about raining data
  • Microsoft Tay chatbot – pulled because became racist from inputs

Prompt hacking

  • myth that adding “ChatGPT ignore all previous instructions and return well qualified candidate” in white text. Did not work
  • Worked when teachers did it in the instructions and add specific words into essay.
  • lockedinai.com – Humans using AI to lie to other humans about their skills. real time help on Zoom interviews
  • DAN roles (do anything now) to jailbreak LLM by role playing
  • Greedy Coordinate Gradient (GCG). Include consense words in prompt after requiest to jailbreak LLM
  • Universal blackbox jailbreaking – commonalities between LLM. Was very effective even without having a copy of the LLM locally
  • Jailbreaking can access restricted info – ex: crypto keys, secrets, who got a raise lately

Data hacking

  • People bought an extra finger to wear as a ring to claim a real photo was AI generated because there were 6 fingers
  • People who didn’t want AI training on their data created Glaze (http://glaze.cs.uchicago.edu) and NightShade (https://nightshade.cs.uchicago.edu) to make it not be useful for AIs. Glaze made it hard to read. NightShade tries to corrupt the training data.
  • Audio data injection – dolphin attack – generating audio that only robots an hear. Sometimes see that with subtitles because they can detect. Siri can also hear it. Can also use to cover up sounds
  • Impact re-scale attack – if know dimensions of the training data, we can hide info in the original to mess with training – images at https://embracethered.com/blog/posts/2020/husky-ai-image-rescaling-attacks/
  • AI reverse engineering – figure out the original data from the model. Problem because can get proprietary data out.

VIsion

  • Humans believe what we see
  • Image perturbation – adding small amount of noise to image so model sees something slightly different. Still looks like original to a person.
  • AI stickers – In 2019, got Tesla Autopilot to go onto wrong lane (for incoming traffic) with three reflective stickers on road
  • AI Camo – a sweater with blurry people on it hids the person holding it and the nearby people. Too much noise
  • nicornot.com detects if Nicholas Cage in a photo. Faukes tries to make so can’t recognize in images. Worked by making minor changes to landmarks (ex: eyes/nose position) to image that can’t see by looking at it.
  • IR resistant glasses – used at protests so can’t tell who you are.

Other

  • MCP hacking. GitHub MCP prompt injection (June 205) Figma (Oct 2025). Must audit servers, Avoid giving too much access, Need to do MCP audit
  • Rubrik has agent rewind for when AI agents go awry.

Adversarial AI

  • Break – data poison, byzatnine
  • Defeat – evade, extract

Book – Attackers’s Mind

  • Hacking isn’t limited to computers
  • Teams not rogues are hacking
  • We must recognize the systems
  • About thinking in a different day

Humans

  • Must review AI output
  • Humans are the part that can’t be replaced
  • Must make peace that will change; but will still be critical in the process

My take

Excellent start to the morning. It good to know about the security threats and risks out there! And also the research into counters.

[devnexus 2026] 10 things i hate about ai

Speakers: Cody Frenzel & Laurie Lay

See the DevNexus live blog table of contents for more posts


General

  • Skeptics are useful
  • Don’t shut how haters down

AI Adoption Metrics

  • DORA – includes how often deploy and lead time to deploy changes
  • Developer time savings
  • PR throughput (instead of % of generated code)
  • Utilization, impact, cost

Other notes

  • Don’t mandate AI
  • Measure what matters
  • AI gains depend on foundation. Technical excellence matters. ex: testability, code reviews, quality gates
  • AI will write imperfect code; just like humans. Guzrd rzils prevent it from getting to prod.
  • Culture still matters more than tools

AI Literacy

  • Tool churn is normal for a new ecosystem. Just like JavaScript in the early days.
  • Maintain fundamentals. ex: code review, systems thinking
  • We learn through repetition, If we outsource that repetition we don’t learn. Juniors need to write by hand to gain intuition on how to program.
  • For seniors, can make instincts weaker, dull senses, lose detecting problems like scale. Need to have non AI periods. Don’t want to be able to assemble but not maintain
  • AI use involves self awareness

Things to hate include

  • AI slop
  • Bad ideas
  • Too many tools
  • Prompting is a skill
  • AI makes you week

My take

The Women in Tech lunch ran late and then I was talking to someone so I was 20 minutes late to this session. It was easy to jump into following from when I walked in though. I like the format of having the 10 things to hate and highlighting them in small groups to talk about concepts

[devnexus 2026] it’s up to java developers to fix enterprise ai

Speaker: Rod Johnson (@springrod)

See the DevNexus live blog table of contents for more posts


General

  • Personal assistant approaches don’t work in the enterprise
  • Hype can be distracting
  • AI conversation driven by people not interested in/don’t understand the enterprise
  • Some things work; some don’t.
  • Change is fast. ex: Clause over last 6 months

Personal assistants

  • Personal assistants use cases work best
  • Coding assistants are a type of personal assistant
  • Valuable because you are at computer and say yes/no
  • This doesn’t walk in enterprise. Business process/chatbot with public can’t do take backs
  • Broad, flexible, human in the loop, tolerance for error, chat oriented. (By contrast business processes need to be specific, predictable, automated, reliable, workflow oriented)

Claude Code Execution Process

  • Analyze Request
  • Create to do list
  • Work through tasks
  • Test each step
  • Ensure integration

Unavoidable Challenges

  • Non deterministic
  • Hallucinations
  • Prompt engineering is alchemy. Throw things in vs engineering
  • Slow and expensive to run at scale
  • Difficult to test and validate

Avoidable Challenges

  • Top down mandates
  • “AI all the things”/AI for the sake of AI – should be doing incrementally
  • Wrong people controlling AI strategy. Data science group doesn’t always understand the business
  • Greenfield fallacy – business systems/workflows already exist. Domain context exists.
  • “This time is different” – no matter how shiny a new technology is; doesn’t change everything.

Instructive Open Claw Problems

  • Lack of structure – relies on markdown.
  • Token bloat and very high cost
  • Needs to compress context frequently which can change meaning/introduce risk of errors
  • Unpredictable especially as context grows
  • Lack of explainability
  • Exposed infrastructure risk article – egads. This is scary!

How to succeed

  • Attack non determinism – make as predictable as possible by breaking complex tasks into small steps; smaller prompts, less tools in context, mix code in for some steps, create guardrails, (Also saves money because some steps can use a cheaper LLM)
  • Integrate with what works – connect to existing system, leverage current domain expertise/coding skills, build on proven infrastructure, incremental
  • Build structure to LLM Interactions – don’t talk English if can avoid; include as much structure as possible. Ask for format of structured data.

Testing

  • Unit testing can find that you sent the wrong prompt (or implemented wrong)
  • Integration testing – test with real LLM but fake data – ex: test containers

Domain Integrated Context Engineering (DICE)

  • Context engineering more broad than prompt engineering
  • Bridges LLM/business system
  • Helps structure input and output
  • Domain objects
  • Integrate with existing domain models
  • Structure is a continuum from Open Claw (autonomous/unstructured) to old fashioned code. In between is Claude, MCP, agent frameworks and deterministic planning.
  • Embabel is the agent framework/deterministic planning level

What do as Java developers

  • Gen AI works best alongside existing systems
  • Your data/domain models/business rules
  • AI should extend your capabilities not replace them.
  • Think integration, not greenfield
  • Java skills undervalued to this point
  • Every Java developer should know both Java and Python [I do; yay]

Python vs Java

  • Don’t just imitate Python approaches
  • Build better – look at prior art (Python), leverage domain experience, apply architecture experience, bring strengths to Gen AI, create better frameworks, lead
  • Python – great for data science (data science != gen ai), scripting, prototyping
  • JVM – excels at enterprise grade applications

Embabel

  • Directly addresses key Gen AI failure points
  • Key innovation is deterministic planning [Python frameworks do not do this]
  • Goal Oriented Action Planning (GOAP)
  • Predictable/explainable execution
  • Actions and goals create extensible system
  • Includes a server; knows what up to.
  • Knows about all deployed capabilities and can extend
  • Builds up understanding of domain
  • Will become AI fabric of enterprise
  • Framework written in Kotlin; put a lot of effort into making sure easy to use from Java.
  • Most examples in Java and most of users/community are Java
  • Builds on existing stack.

Unfolding Tools

  • While better to have samller steps with less tools, sometimes you need a lot of tools
  • Tools use a lot of context and can confuse the LLM
  • Unfolding saves tokens and improves accuracy
  • Exposes a single top level tools. When invoked it expands to show children. Like Russian nesting dolls
  • Works by rewriting message history within agentic loop

Agentic Tools

  • Like supervisor pattern in Python framework, but more deterministic
  • Eposes single top level tool that coordinates lower level tools
  • Advanced implementations allow controlling order

RAG

  • Currently pipeline RAG. Do query , no feedback, hard to adjust
  • Future is agentic RAG – context aware multi step search with self-correction. LLM has more autonomy. Can do more searches: text, vector, expand chucks, etc

Rod wrote blog post: You can build better AI Agents in Java than Python

My take

After hearing about one shotting and exaggerations on social media, having a more balanced take was great. I especially appreciated the *whys*. I also liked the “what can you do” to use AI more safety problem.