(I'm one of the founders of the Recurse Center.)<p>At least so far, the impact of LLMs has been overall positive for RC. We have a big-tent community, which includes everyone from those who avoid LLMs at all costs to those who are all-in on them, and everyone in between. We have folks who come to our retreats with the explicit goal of <i>not</i> programming with LLMs[1] (usually because they've been using them extensively at their previous jobs) and others who use RC as an opportunity to learn how to program effectively with AI. There's also a lot of discussion here about how (and to what extent) to use LLMs effectively in the context of learning.<p>We wrote up our position on AI last July, and while a lot has changed since then, I think all our conclusions still hold up well[2]: "...whether you choose to embrace or avoid AI in your work at RC, you will need to build your own mental structures to grow as a programmer. When using AI, use it to amplify your ambitions, not to abdicate your agency. And regardless of what you do, be curious about and kind to the people around you."<p>Like everyone else, I have no idea what the future holds, but I'm confident we'll find a way to navigate it. I'm sure I have some motivated reasoning here, but I really do believe that humans will still want to understand and build things no matter how good the models get.<p>[1] <a href="https://miguelconner.substack.com/p/im-coding-by-hand" rel="nofollow">https://miguelconner.substack.com/p/im-coding-by-hand</a><p>[2] <a href="https://www.recurse.com/blog/191-developing-our-position-on-ai" rel="nofollow">https://www.recurse.com/blog/191-developing-our-position-on-...</a>