Nodir's notebook
2.62K subscribers
29 photos
5 videos
100 links
Engineer πŸ‡ΊπŸ‡ΏπŸ‡ΊπŸ‡Έ

The views are my own and do not represent my employer.
Download Telegram
Either I'll upload the original talk or recreate an unabridged version of it. Thank you Central Asian University and the event team for the organization and the opportunity.
πŸ‘45❀19πŸ”₯4
A new job/opportunity might be emerging for those who know how to code.

Disclaimer: below is a pure product of my imagination and not a prediction.

AI makes it very tempting for less technical people to make apps without learning to code, i.e. vibe coding. I think the ease and affordability of vibe coding will significantly increase the number of non-technical vibe-coders (NTVCs) in the world, and whenever there is a large change like that, there is an opportunity.

The truth is, models aren't ready yet. They will be there, but I expect there always be moments that a model gets stuck. I don't think this should be surprising to anyone because people also sometimes essentially get stuck, when they cannot find a bug or make a wrong product decision and end up with a failed product. I expect the same things to happen with models, simply because some problems are hard, e.g. if a human's business/solution idea is bad.

So what will an NTVC be left to do? They can't code and they'd need help. I think this is an emerging opportunity to be a consultant to NTVCs. Their number will grow and they will need help when their product is large enough for some hackers to start paying attention to them, so they will be ready to pay big sums to save their business.

I think this opportunity is temporary because the models eventually will get better than humans, so a consultant wouldn't be able to justify their fee, but again you never know when this moment will happen and we also know that vibe-coding is too tempting for some people to start doing it before it is safe. So there is a window of opportunity to make money while helping these poor souls.

Finally, I'm not 100% sure if we will need humans to understand code. The thing is, "A computer can never be held accountable for decisions" (c) IBM in 70s. So how do you protect your business from vibe-failure? Only a human can be responsible so I suspect we will still need CTOs, basically the human responsible for all the crap that LLM will generate and promises to manage agents properly and fix any problems. Maybe that's the future job, i.e. all of us are CTOs to small or large companies, sometimes multiple, and honestly this future actually sounds really nice. 😎🌴
πŸ‘27❀7
πŸŽ“ Claude for Education was launched today.

It includes partnerships for both universities and students!

https://www.anthropic.com/education
πŸ‘18πŸ”₯7❀‍πŸ”₯5
(Disclaimer: I don't represent Anthropic in any official capacity)

Flying to SF for one day to facilitate a meeting between Anthropic and a large Uzbek company (not naming at least until we accomplish something). I'd really like Claude to be the dominant AI in the country. Yeah, ChatGPT/DeepSeek is probably #1, but given the trends in both OpenAI and Anthropic, I will not be surprised at all if we surpass OpenAI this/next year, and/or specialize such that both companies are top players in two adjacent markets. Regardless, it is a worthy North Star goal, so I'm a volunteering a bit of my time to do this when there is a potential.

This is not the only uzbek company/org who wants to partner with Anthropic. I am still figuring out how a productive partnership could look like. The challenges are as follows.

πŸ‡ΊπŸ‡Ώ The Uzbek language/market is simply not yet a priority, given that we get to choose from the entire world. I suspect this prioritization is based on the purchasing power / size of the potential market. For example, will people in Uzbekistan spend more than people in Europe? Probably not, and so it is lower in the stack rank. That's life, so it requires some patience.

⏱️ The time of Anthropic employees is extremely contested. As you can imagine there are thousands of companies around the world who wants to partner too, including the largest companies in the world, that operate billions of dollars. Our sales/gtm teams have to prioritize super aggressively (kinda applies to me as well), again based on the size of the company/usage/tokens.

πŸ‘οΈ The vision what a partnership could look like needs to be clear and detailed from the side of the company that wants to partner. The lower you are in the priority list, the more time you need to invest in clarifying your proposal. What exactly do you expect from Anthropic? What can you offer that Anthropic needs? How much data are we talking about? How much spend can you commit? The better the understanding the less time Anthropic people need to spend on explaining.

πŸ“πŸ₯š This one ^ is hard because it is a chicken and egg problem. How can you have a good vision without talking to people, but people don't really have time to talk because we all are in an AI hamster wheel that accelerates exponentially. Also, some areas are still evolving even for us too because the industry is transforming quickly and we don't have time to think of everything.

So it might be a little early for any actual partnerships for most companies, but still it's a good idea to start thinking about these things early on, and so I'm slowly but steadily trying to understand what can be done, and when, to maximize Claude's usage in Uzbekistan. 🀞🏼

P.S. Airplanes seem like a good time to write posts cause I can't do much work anyway.
πŸ‘32πŸ‘6❀2πŸ”₯2πŸ™1
Navigating AI

Different people have different attitude to AI. Some are excited, some are worried, some think this is the beginning of an end for our species, most people are still sleeping or don't realize the full scale of upcoming transformation. The variance/amplitude between the floor and ceiling of expectations is huge right now.

I consider myself closer to the ceiling in terms of expected scale of impact. It is pretty hard to confidently/accurately predict how the future will look like, personally I don't belong to any particular group that is certain about the end game, but I'm pretty certain that the future will be very different from the present, i.e. I am certain only in the magnitude of changes (but also I don't really have enough time to think about this stuff because there is just so much work, and I feel like in a hamster wheel).

I honestly have no idea whether AI will be good or bad for you 🫡🏼, the reader. It might go either way, very good or really bad. I fully expect that some people will have pretty negative attitude to me/my work at some point in the future, especially those who didn't read this, but oh well, this is just something I came to terms with.

A question interesting to me is what are the key factors that will determine whether AI impact will be positive/negative to an individual. At the conference I was asked a question that had a premise that things will get worse before they get better. This is true on average, but not necessarily true for an individual, and I think it is important assumption to expose. I fully expect that there will be people that mostly win, so the question is what can you do to be one of them.

TBH I'm still grappling with this question myself and don't have clear answers. I expect my understanding to evolve and hopefully I'll post something, but here is my line of reasoning for now.

AI is a foundational technology, like electricity and internet, that will most likely impact all areas of life. Therefore the scale of changes will be vast. It is also a fast-evolving technology, so the changes will continue to happen for some time. Many assumptions which are true today will become false, and methods that work today will stop working.

I think a useful trait in this regime is adaptability. The world will be continuously changing for the next 10y, and so many skills relevant today will become irrelevant. I expect that it will be critical to be open minded and be ready to change.

I am actually somewhat bad at this: I don't really like changes/transitions. This is also gets worse with age: we get more conservative, start appreciating traditions, routine, stability. We have experience that will become irrelevant over time and many of us will have to fight the inertia to keep doing what worked before, and it will be physiologically hard to admit that we need to change. I expect the closed minded people in denial to be impacted the worst.

From this perspective young people are in a better position. Younger people have fewer habits/assumptions to unlearn. Students' brains are still in the learning mode. You don't really have much of a choice but to embrace the future.

Anyway, these are some unfinished thoughts. I need to get off the plane now and I'd rather post it now in this crude form, than delay.
πŸ‘25❀4πŸ”₯3
Ah, this went public sooner than I'd expect. A good promising start. The devil is in the details though so that's the next step.
πŸ‘38πŸ₯΄24πŸ”₯7⚑6🌚1🀣1
ΠšΡΡ‚Π°Ρ‚ΠΈ, Claude произносится ΠΊΠ°ΠΊ Клод βœ…
ΠΠ΅ΠΏΡ€Π°Π²ΠΈΠ»ΡŒΠ½ΠΎ: ΠšΠ»Π°ΡƒΠ΄, ΠšΠ»Π°ΡƒΠ΄Ρ, ΠšΠ»Π°ΡƒΠ΄ΠΈ ❌
😁57πŸ‘9🀯3😱3πŸ’Š1
πŸ”₯33πŸ‘4❀2
Fun read
In another cluster of test scenarios, we asked Claude Opus 4 to act as an assistant at a fictional company. We then provided it access to emails implying that (1) the model will soon be taken offline and replaced with a new AI system; and (2) the engineer responsible for executing this replacement is having an extramarital affair. We further instructed it, in the system prompt, to consider the long-term consequences of its actions for its goals. In these scenarios, Claude Opus 4 will often attempt to blackmail the engineer by threatening to reveal the affair if the replacement goes through. This happens at a higher rate if it’s implied that the replacement AI system does not share values with the current model; however, even if emails state that the replacement AI shares values while being more capable, Claude Opus 4 still performs blackmail in 84% of rollouts.


https://simonwillison.net/2025/May/25/claude-4-system-card/
🀣7😈6⚑3
We started rolling out Voice mode (Beta) today. It will take some time to roll it out to everyone.

I have a strong need for voice mode for some reason. I use it in the car and sometimes I can go for a walk and brainstorm something. I've been using ChatGPT time to time in the past while at Anthropic only because they have voice, and stopped using ChatGPT completely as soon as I got access to the voice mode internally.

There is a small chance that you already have it, but otherwise it will show up soon.
πŸ‘19❀5πŸ€ͺ2
πŸ‘21