(评论)
(comments)

原始链接: https://news.ycombinator.com/item?id=38190401

第二次世界大战中的反潜战为人工智能技术的现代发展提供了有趣的历史视角。 虽然战争期间的潜艇是由船员操作的,但技术的进步提高了通信能力和战略决策能力,为水下船舶的进一步技术创新铺平了道路。 随着人工智能技术的不断进步,它提高了将这些先进技术融入现代潜艇的可能性,以增强其在未来冲突中的能力和有效性。 尽管人们担心潜在的网络安全漏洞和确保人工智能系统的可维护性,但仍需要在这些领域继续进行研究和开发工作,以确保技术优势完好无损。 最终,包括人工智能在内的先进技术的整合,为改善陆地、海洋和空中领域的军事行动提供了机会,从而增强态势感知和战术灵活性。

相关文章

原文
Hacker News new | past | comments | ask | show | jobs | submit login
Major outages across ChatGPT and API (openai.com)
484 points by d99kris 17 hours ago | hide | past | favorite | 498 comments










I used Google Bard for the first time today specifically because ChatGPT was down. It was honestly perfectly fine, but it has a slightly different tone than ChatGPT that's kind of hard to explain.


People constantly recommend Bard to me, but it returns false and/or misleading info almost every time


I had the opposite experience. I was trying to figure out what this weird lightbulb was that I had to replace, and it had no writing on it. I uploaded the description and picture to gpt4, and it just got it clearly wrong over and over. Tried bard, got it right in the first try, with links to the product. I was extremely impressed.


I had a similar experience with Google lens recently. I've gotten used to Yandex image search being better than Google's for many searches, but I needed to figure out what model of faucet I had on my sink, and Google nailed it. My hunch is that because of all the work they did on Google shopping gathering labeled product images and things like that, they have an excellent internal data set when it comes to things like your light bulb and my sink.


The vast knowledge trove of Google can't be understated, even if sometimes the model isn't as competent at certain tasks as OpenAI's GPT models.


Isn't there just a comment today on HN saying Google had an institutional reluctance to use certain data sets like libgen? I honestly don't think Google used everything they had to train their LLM.

https://news.ycombinator.com/item?id=38194107



If there's one thing that's becoming clear in the open source LLM world, it's that the dataset really is the 'secret sauce' for LLMs. There are endless combinations of various datasets plus foundation model plus training approach, and by far the key determinant of end model performance seems to be the dataset used.


> it's that the dataset really is the 'secret sauce'

alwayshasbeen.jpg

There have been articles about how "data is the new oil" for a couple of decades now, with the first reference I could find being from British mathematician Clive Humby in 2006 [0]. The fact that it rings even more true in the age of LLMs is simply just another transformation of the fundamental data underneath.

[0] https://en.wikipedia.org/wiki/Clive_Humby#cite_ref-10



This is almost certainly just delegating to Google Lens, which indeed works great.

Bard's probably just a middle man here.



I would think that ChatGPT is also delegating to some other subsystem.


Right, a glance at the new Assistant API docs, which seems to mirror ChatGPT in functionality, suggests that the "Assistant" determines which tool to use, or which models to use (code or chat) to generate a response message. The API is limited in that it can't use Vision or generate images, but I imagine that those are just "Tools" too that the assistant has access to.


with a poorer dataset than google


Unless wired to Bing, probably not?


Ah that’s interesting, did know bard had multi modal inputs.


I asked Bard the question "Who is Elon Musk?"

The response: "I'm a text-based AI, and that is outside of my capabilities."



That's interesting. Did you have anything else in the thread, or was it just the first question?

For me it returns a seemingly accurate answer [1], albeit missing his involvement with Twitter/X. But LLMs are intrinsically stochastic, so YMMV.

[1] https://g.co/bard/share/378c65b56aea



It was the first question in the thread, and I've been testing queries along these lines on it for a while. Interestingly, it started writing a response initially and then replaced it with that. It used to refuse to answer who Donald Trump is as well, but it seems that one has been fixed.

Another interesting line of inquiry (potentially revealing some biases) is to ask it whether someone is a supervillain.



So Elon is the engine behind Bard?


What makes you think ChatGPT isn't also returning false and/or misleading info? Maybe you just haven't noticed...

Personally, I struggle with anything even slightly technical from all of the current LLM's. You really have to know enough about the topic to detect BS when you see it... which is a significant problem for those using it as a learning tool.



True, it often returns solutions that may work but are illogical. Or solutions that use tutorial style code and fall apart once you tinker a bit with it.


This is my problem with chatgpt and why I won't use it; I've seen it confidently return incorrect information enough times that I just cannot trust it.


I get information from unreliable sources all the time. In fact all my sources are unreliable.

I just ignore how confident ChatGPT sounds.



The version with search will give you links to the references it bases its answers on.


You still have to go read the references and comprehend the material to determine if the GPT answer was correct or not.

I don't know the name for the effect, but it's similar to when you listen/watch the news. When the news is about a topic you know an awful lot about, it's plainly obvious how wrong they are. Yet... when you know little about the topic, you just trust what you hear even though they're as likely to be wrong about that topic as well.

The problem is people (myself included) try to use GPT as a guided research/learning tool, but it's filled with constant BS. When you don't know much about the topic, you're not going to understand what is BS and what is not.



In my particular case, the fact that it returns bullshit is kind of useful.

Obviously they need to fix that for realistic usage, but I use it as a studying technique. Usually when I ask it to give me some detailed information about stuff that I know a bit about, it will get some details about it wrong. Then I will argue with it until it admits that it was mistaken.

Why is this useful? Because it gets "just close enough to right" that it can be an excellent study technique. It forces me to think about why it's wrong, how to explain why it's wrong, and how to utilize research papers to get a better understanding.



The Gell-Mann Amnesia Effect


Gpt-4 has the lowest hallucination rates:

> OpenAI’s technologies had the lowest rate, around 3 percent. Systems from Meta, which owns Facebook and Instagram, hovered around 5 percent. The Claude 2 system offered by Anthropic, an OpenAI rival also based in San Francisco, topped 8 percent. A Google system, Palm chat, had the highest rate at 27 percent.

https://www.nytimes.com/2023/11/06/technology/chatbots-hallu...



i just skimmed through the article - it seems that the numbers are quoted from a company called Vectara. Would be interesting to see how they are getting to this estimate


Sounds like Bard is catching up with ChatGPT in features then.


I noticed that too, it just doesn't seem that good


Bard has different training data and regime, that alone is enough to start to understand why they are different.

The main thing as a user is that they require different nudges to get the answer you are after out of them, i.e. different ways of asking or prompt eng'n



Yeah, which is why I use the paid version of ChatGPT still, instead of the free Google Bard or Bing AI; I've gotten good enough at coercing the GPT-4 model to give me the stuff I want.

Honestly, $20/month is pretty cheap in my case; I feel like I definitely extract much more than $20 out of it every month, if only on the number of example stubs it gives me alone.



I stopped paying OpenAI because they went down or were "too busy" so much of the time I wanted to use it. Bard (or more so the VertexAI APIs) are always up and reliable and do not require a monthly fee, just the per call


How often was OpenAI down that you wanted to look elsewhere?


Not the person you asked, but (paid) ChatGPT was down so often for me that I almost cancelled... Until I switched to connecting via VPN. Only one outage since then For some reason whole swaths of the Spectrum IP block in Gilroy has trouble.


The custom prompts feature, and the "about me" mostly fixed this for me, with my usual conversations. In both, I convinced it that I'm competent, and that it doesn't need to hold my hand, so much.

I've also noticed that API based clients (rather than the web or iOS client) result in conversations that hold my hand less. The voice client seems hopeless though, probably because I write ok, but have trouble saying what I want before the stupid thing cuts me off. It seems to love making lists, and ignoring what I want.



Is there a difference between OpenAI and Bing AI? They both claim to use GPT-4.


I've found that Bard has an overly aggressive filter, like if I'm brainstorming ideas about thieves in a fantasy world (think Lies of Locke Lamora), it will frequently refuse to cooperate.

I think it's running some kind of heuristic on the output before passing it to the user, because slightly different prompts will sometimes succeed.

ChatGPT's system is smart enough to recognize that fantasy crimes are not serious information about committing real crimes or whatever.



I used it for the first time today too, for the same reason. It was slower and much worse at coding. I was just asking it for SQL aggregation queries and it just ignored some of my requirements for the query.


In my case, I was just asking it for a cheeky name for a talk I want to give in a few months. The suggestions it gave were of comparable quality to what I think ChatGPT would have given me.


In my experience, Bard is much, much better at creative things than at things that have correct and incorrect answers.


I subscribe to the belief that, for a chat model with the same parameters, creativity will be proportional to tendency to hallucinate, and inversely proportional to the factual answers. I suspect an unaligned model, without RLHF, wouldn't adhere to this.


I have seen noticably worse results with Bard, especially with long prompts. Claude (by Anthropic) has been my backup to ChatGPT.


ChatGPT was down, so of course it'd be slower. And possibly that accounts for some quality loss as well.

For a fair comparison, you probably need to try while ChatGPT is working.



> ChatGPT was down, so of course it'd be slower

Actually at Google scale I wouldn't expect so



Yeah but Google missed the boat when it came to hardware accelerators specifically meant for LLMs (their proprietary TPUs aren't optimized for LLMs) so it's just a matter of whether Google or Microsoft paid Nvidia more. In the current cost cutting climate at Google I doubt the answer is so certain.


The extension with gpt4 as a backend was ime extremely slow as standard. I've not tried it again with the v7 model though which is supposed to be a lot faster


I used https://you.com/chat wasn't bad, they have a free month trial coupon "codegpt" for the GPT4 model and GPT3.5 is free ...


Phind is pretty good for coding (LLama 2 trained on billions of extra code tokens) and is still up https://www.phind.com/s


I am curious how people are using Phind.

I actually had a discussion with Phind itself recently, in which I said that in order to help me, it seems like it would need to ingest my codebase so that it understands what I am talking about. Without knowing my various models, etc, I don't see how it could write anything but the most trivial functions.

It responded that, yes, it would need to ingest my codebase, but it couldn't.

It was fairly articulate and seemed to understand what I was saying.

So, how do people get value out of Phind? I just don't see how it can help with any case where your function takes or returns a non-trivial class as a parameter. And if can't do that, what is the point?



I am using Phind quite a lot. It's using it's own model along GPT 4 while still being free.

It is also capable to perform searches, which lead me - forgive me founders - to abuse it quite a lot: whenever I am not finding a good answer from other search engines I turn up to Phind even for things totally unrelated to software development, and it usually goes very well.

Sometimes I even ask it to summarize a post, or tell me what HN is talking about today.

I am very happy with it and hope so much it gains traction!



Founder here. We have a VS Code extension that automatically fetches context from your codebase.


Does the plugin use the codebase code in the model itself for training, or does it stay locally? Just curious about the privacy.


We won’t use your codebase for training. And there is an option you can toggle that disables data retention entirely.


So if source available libraries are imported, they are also parsed by the AI?

So if I create a GPT for my open-source library as a way to fund it, all these copilot etc. are going to compete with me?

Just wondering because that would be a bummer to not have this avenue to fund open-source code.



I am not related to Phind or any other AI company, but yes, this is definitely the case, and you should assume that they will be ingesting your code through regular web scrapes now (giving extremely general knowledge about your library) and through reading specifically the library source code soon (this is what you are asking about here). If you wanted to try this strategy, I would suggest that you do it by providing the model with a large database of high-quality examples specific to your library (so, perhaps the examples section of your website, plus snippets from open source projects that use the library). These will probably be the last to be specifically ingested by general coding models.


Woah , didn't know that! Thanks for pointing out!


Thanks for releasing Phind-CodeLLaMA-34B-v2, it's been helping me get up to speed with node and web apps and so far it's been spot on. :) Super impressive work.


I've used it to interpret messages in the Mac console app. I wish it could explain what causes certain messages to appear over and over.


I had great results with Phind. Their newest fine tune model (V7) has been a pleasant experience and better than most open source models out there.

Nit: your link has a trailing "s" which makes it 404 :)



Me too. For past few weeks, I had been working on my AHK scripting with Phind. It produced working code consistently and provided excellent command line for various software.

Also I use it for LaTeX, too. It is very helpful providing various package than trying to hunt more information through Google. I got a working tex file within 15 min than it took me 3 weeks 5 years ago!



I’ve had some consistency issues with phind but as a whole I have no real complaints, just glitches here and there with large prompts not triggering responses and reply options disappearing.

As a whole I think it works well in tandem with ChatGPT to bounce ideas or get alternate perspectives.

(I also love the annotation feature where it shows the websites that it pulled the information from, very well done)



Been playing with Phind for a while and my conclusion is: the Phind model works well on those long existing stuff like C++ libraries, but works generally bad on newer stuff, such as composing LCEL chains.


Phind seems to be down

"The inference service may be temporarily unavailable - we have alerts for this and will be fixing it soon."



The first coding question I tested it on, it gave me something completely wrong and it was pretty easy stuff, I’m sure it gets a lot right but this just shows unreliability


Phind for me has worked pretty bleh compared to just back and forth conversation with a python GPT4 bot I made lol.


GPT4 completion or chat model?


To use the api and stop them logging your chat? Have you compared to aider? Also got it on a repo?


is this a shameless plug or an honest referral


> 404 This page could not be found.

Is /s a self-fulfilling sarcasm indicator or a typo?



Holy hell, was shitting bricks, considering I JUST migrated most services to Azure OpenAI (unaffected by outage) — right before our launch about 48 hours back. What a relief.


congrats!


I was gonna say that this is Bard's chance to shine, but it looks like Bard is also having an outage!


I went to use Bard, and it looks so clean, such a nice UI. And the response looks so well organized, simply beautiful. If the AI only were as good as OpenAI's...


I was going to say that we need to grab our tin foil hats, this can't be an coincidence :D


"Something went wrong. Sorry, Bard is still experimental". Chance wasted I guess.


I can't believe it's still not available in canada


They started talking to each other.


And that's one picket line that nobody is going to cross.


Remember when all the niche MMOs would lag out on Tuesday nights as soon as the WoW servers went down?


Claude?


Bard is still playing scared, it isn’t even international yet


Bard is available in 40 languages and 230+ countries: https://support.google.com/bard/answer/13575153?hl=en


I don’t think there are 230+ countries, at least not on this planet.


I'm in the UK and it works fine. I'm assuming people would be throwing a fit if it didn't work in the US, so clearly it is international.


Bard devs secretly built it on top of OpenAI's api? /s


More likely OpenAI is using GCP or some other service that Bard is also using?


more like people are rushing to bard since they cant use chatgpt, causing a huge spike


I doubt it - Microsoft wants OpenAI on Azure 100%


I noticed the outage. It feels like a lot of people use it like training wheels on a bicycle until they forget how to ride without it.


Seems more like a crutch in that sense, then. If you can’t ride without it, it hasn’t trained you anything.


i used bard today. it's gotten a lot better.


I hope so!

I'm still surprised by the problems with it. Last month it lied about some facts then claimed to have sent an email when asked for more details.[1]

Then apologized for claiming to send an email since it definitely did not and "knew" it could not.

It's like a friend who can't say 'I don't know' and just lies instead.

1. I was asking if the 'Christ the King' statue in Lisbon ever had a market in it, a rumor told to me by a local. It did not, contrary to Bard's belief.



Bard promised me it would design a website for me. It said it’d get back to me in a couple of weeks. I can’t even remember the prompt but it was basically LARPing as a Wordpress theme designer.


At least we know what Google has been using our Gmail data for!


Well, don't keep us in suspense. Did you ever hear back?

:)



Fortunately for OpenAI, they have no SLAs: https://help.openai.com/en/articles/5008641-is-there-an-sla-...


I say this as a huge fan of GPT, but it's amazing to me how terrible of a company OpenAI is and how quickly we've all latched onto their absolutely terrible platform.

I had a bug that wouldn't let me login to my work OpenAI account at my new job 9 months ago. It took them 6 months to respond to my support request and they gave me a generic copy/paste answer that had nothing to do with my problem. We spend tons and tons of money with them and we could not get anyone to respond or get on a phone. I had to ask my coworkers to generate keys for everything. One day, about 8 months later, it just started working again out of nowhere.

We switched to Azure OpenAI Service right after that because OpenAI's platform is just so atrociously bad for any serious enterprise to work with.



I've personally never scaled a B2B&C company from 0 to over 1 billion users in less than a year, but I do feel like it's probably pretty hard. Especially setting up something like a good support organization in a time of massive labor shortages seems like it would be pretty tough.

I know they have money, but money isn't a magic wand for creating people. They could've also kept it a limited beta for much longer, but that would've killed their growth velocity.

So here is a great product that provides no SLA at all. And we all accept it, because having it most of the time is still better than having it not at all ever.



I'm not judging them at all as I agree with your core statement, just saying it's quite remarkable that companies around the world who spend 6 months on MSA revisions in legal over nothing are now OK with a platform that takes 6 months to respond to support requests.


I wonder if they spend time trying to do support via GPT4 itself.


GPT-4 would be more responsive. They ignore support requests for weeks unless you keep reminding them.


The remarkable part isn't that OpenAI sucks - it's that people use it anyway.


> I say this as a huge fan of GPT, but it's amazing to me how terrible of a company OpenAI is and how quickly we've all latched onto their absolutely terrible platform.

Your example is clearly not acceptable, but I can see reasons for it.

OpenAI apparently was somewhere between "I can't see people finding this useful" and "I guess" when deciding on releasing ChatGPT at all in the first place.

If that's the case, I doubt they were envisioning a flood of users, who needed a customer support person to handle their case. They have to spin-up an entire division to handle all of this. And I'm sure some of the use-cases are going to get into complex technical issues that might be hard to train people for.

They can no longer remain a heads-down company full of engineers working on AI.

I'm not excusing it, but I can see why things like your situation might occur. Although 6 months for a response is obviously ridiculous. If you are paying them a significant amount of money, and it is impacting your business, then that's all on OpenAI to fix ASAP.



ChatGPT has been broken for me for two months, regardless of whether I use the iOS app or the web app. The backend is giving HTTP 500 errors – clearly a problem on their end. Yet in two months I haven’t been able to get past their first line of support. They keep giving me autogenerated responses telling me to do things like clear my cache, turn off ad blockers, and provide information I’ve already given them. They routinely ignore me for weeks at a time. And they continue to bill me. I see no evidence this technical fault has made it to anybody who could do anything about it and I’m not convinced an actual human has seen my messages.


> I had a bug that wouldn't let me login to my work OpenAI account at my new job 9 months ago.

I also cannot login on Firefox (latest version) with strict privacy settings and AdNauseam on desktop.. and a few weeks ago they broke their website on iOS v14 as well for no apparent reason (it certainly didn't make me to download their app since that require v16.1+).



Let's not forget the most basic problem with OpenAI: It isn't open.


> newspeak, propagandistic language that is characterized by euphemism, circumlocution, and the inversion of customary meanings.

1984 got so many things so right about the future.



OpenAI is relatively young on the adoption and scaling front.

Also, they need to remain flexible most likely in their infrastructure to make the changes.

As an architecture guy, I sense when the rate of change slows down more SLA type stuff will come up, or may be available first to Enterprise customers who will pay for the entire cost of it. Maybe over time there will be enough slack there to extend some SLA to general API users.

In the meantime, monitoring API's ourselves isn't that crazy. Great idea to use more than one service.



I'd be curious to hear about the workflows people have come up with using ChatGPT. I'm still in the realm of "I don't know how to do this" or "I forgot the exact incantation to that" or "is the an X that does Y in framework Z?"


I can share one set that we have.

Basically, we use AI to do a lot of formatting for our manuals. It's most useful with the backend XML markups, not WYSIWYG editors.

So, we take the inputs from engineers and other stakeholders, essentially in email formats. Then we pass it through prompts that we've been working on for a while. Then it'll output working XML that we can use with a tad bit of clean-up (though that's been decreasing).

It's a lot more complicated than just that, of course, but that's the basics.

Also, it's been really nice to see these chat based AIs helping others code. Some of the manuals team is essentially illiterate when it comes to code. This time last year, they were at best able to use excel. Now, with the AIs, they're writing Python code of moderate complexity to do tasks for themselves and the team. None of it is by any means 'good' coding, it's total hacks. But it's really nice to see them come up to speed and get things done. To see the magic of coding manifest itself in, for example, 50 year old copy editors that never thought they were smart enough. The hand-holding nature of these AIs is just what they needed to make the jump.



Did you have any scripts or other explicit “rules-based” systems to do this before? Is it a young company?

It sounds like a pretty old and common use case in technical writing and one that many organizations already optimized plenty well: you coach contributors to aim towards a normal format in their email and you maintain some simple tooling to massage common mistakes towards that normal.

What prompted you to use an LLM for this instead of something more traditional? Hype? Unfamiliarity with other techniques? Being a new company and seeing this as a more compelling place to start? Something else?



GPT-4 is quite capable of writing function-length sections of code based only on descriptions. Either in a context where you're not sure what the a good approach is (for myself, when writing Javascript for example), or when you know what needs to be done but it's just somewhat tedious.

Here's a session from me working on a side project yesterday:

https://chat.openai.com/share/a6928c16-1c18-4c08-ae02-82538d...

The most impressive thing I think starts in the middle:

* I paste in some SQL tables and the golang structrues I wanted stuff to go into, and described in words what I wanted; and it generated a multi-level query with several joins, and then some post-processing in golang to put it into the form I'd asked for.

* I say, "if you do X, you can use slices instead of a map", and it rewrites the post-processing to use slices instead of a map

* I say, "Can you rewrite the query in goqu, using these constants?" and it does.

I didn't take a record of it, but a few months ago I was doing some data analysis, and I pasted in a quite complex SQL query I'd written a year earlier (the last time I was doing this analysis), and said "Can you modify it to group all rows less than 1% of the total into a single row labelled 'Other'?" And the resulting query worked out of the box.

It's basically like having a coding minion.

Once there's a better interface for accessing and modifying your local files / buffers, I'm sure it will become even more useful.

EDIT: Oh, and Monday I asked, "This query is super slow; can you think of a way to make it faster?" And it said, "Query looks fine; do you have indexes on X Y and Z columns of the various tables?" I said, "No; can you write me SQL to add those indexes?" Then ran the SQL to create indexes, and the query went from taking >10 minutes to taking 2 seconds.

(As you can tell, I'm neither a web dev nor a database dev...)



This lines up with my general experience with it. It’s quite proficient at turning a decently detailed description into code if I give it the guard rails. I’ve compared it to having a junior developer at your disposal. They could do a lot of damage if they were given prod access but can come back with some surprisingly good results.


Are you at all worried about what happens if we have a generation of human junior developers who just delegate to this artificial junior developer?

I do. If too many of our apprentices don’t actually learn how to work the forge, how ready will they be to take over as masters someday themselves?

I can see how ChatGPT was useful to the grandparent today, but got very disturbed by what it might portend for tomorrow. Not because of job loss and automation, like many people worry, but because of spoiled training and practice opportunities.

I liked your take, so I’d be curious to hear what you think.



> what happens if we have a generation of human junior developers who just delegate to this artificial junior developer?

Looking forward to the Y2K levels of highly paid consulting work becoming available now you mention it



chatgpt doesn't just program, is interactive, this will make junior dev. more emphasized in their strength and not, while gaining a lot of experience


We're eventually going to have to give up on the notion that we must understand the inner workings of the things we build. That's arguably starting to happen now. Not 100% sure it's a bad thing, but it's certainly scary.

We've long since reached the point at which no one can be said to be a true polymath ( https://en.wikipedia.org/wiki/The_Last_Man_Who_Knew_Everythi... ). Having lost the ability as individuals to know something about everything, we're now losing the ability to know everything about anything.

I'm pretty sure that while the most popular programming languages today are Python and Javascript, the most popular ones 10 years from now will be English and Mandarin. Everything we know about software development is about to change. It's about time.



Wow, so, you're not a DBE or DBA but are applying indexes across a database without concern because...a computer model spat out that you should?


This is a local SQLite database into which I had slurped a load of information about git commits to do data analysis. If I'd somehow destroyed the database, I would have just created it again.


I am also not those. Is there something wrong with too many indexes, assuming memory isn't a huge constraint?


Biggest one is write overhead: indexes have to be updated every time a record is created or an indexed column is updated. This must be done within the same transaction as the create or update, so you're adding an unnecessary overhead to every single one of those ops. Now, if the data is relatively small or the use-case doesn't warrant it, it doesn't matter.

Lesser issues: additional strain on index rebuilding whenever that happens; messing with execution plans and causing the query planner to be inefficient; primary/secondary memory overhead; or if your DB engine uses locks you can run into a myriad of issues there.

I'm all for SWEs learning about databases, as I'm morally opposed to the proliferation of ORMs and the like, but I don't think ChatGPT is the right way to go about things long-term. It's similar to Googling or using StackOverflow: yes, you will find information that is relevant to what interests you at the moment, but it's soon forgotten and does nothing to help build long-term mental models.



I like to use it for one-off scripts. For example, I downloaded a bunch of bank statements the other day, and they had a format something like, "Statement-May-1-2023.pdf" and I asked GPT for a powershell script to convert that to "2023-05-01-BankName-Statement.pdf"

It saved a bunch of manual work on a throwaway script. In the past, I might have done something in Python, since I'm more familiar with it than powershell. Or, I'd say, "well, it's only 20 files. I'll just do it manually." The GPT script worked on the first try, and I just threw it away at the end.



This is what convinced me to pay. I'm completely fine with spending $1/workday if it saves me 10 to 15 minutes per day of bore.


I experimented with using a streamlined workflow to use ChatGPT (GPT-4) for coding tasks.

In the end I settled on a standalone desktop app to "compose" prompt with source code, instructions and formatting options which I can just copy paste into ChatGPT.

The app is available for download if anyone is interested: https://prompt.16x.engineer/



ChatGPT is a good editor for the papers I write for school. Even for short sentences I don't like, I'll ask it for some options to reword/simplify.

I also use it heavily for formatting adjustments. Instead of hand-formatting a transcript I pull from YouTube, I paste it into Claude and have it reformat the transcript into something more like paragraphs. Many otherwise tedious reformatting tasks can be simplified with an LLM.

I also will get an LLM to develop flashcards for a given set of notes to drill on, which is nice, though I usually have to heavily edit the output to include everything I think I should study.

In class, if I'm falling behind on notetaking, I'll get the LLM to generate the note I'm trying to write down by just asking it a basic question, like: "What is anarchism in a sentence?" That way I can focus on what the teacher is saying while the LLM keeps my notes relevant. I'll skim what it generates and edit to fit what my prof said, but it's nice because I can pay better attention than if I feel I have to keep track of what the prof might test me on. This actually is a note-taking technique I've learned about where you only write down the question and look up the answer later, but I think it's nice I now can do the lookup right there and tailor it to exactly how the prof is phrasing it/what they're focusing on about the topic.



I don't know what you do for a living/hobby, or what you might be interested in using ChatGPT to do for you, but here is how I became familiar with it and integrated it into my workflow. (actually, this is true for regular copilot too)

What I'm about to say is in the context of programming. I have the tendency to get caught up in some trivial functionality, thus losing focus on the overall larger and greater objective.

If I need to create some trivial functionality, I start with unit tests and a stubbed out function (defining the shape of the input). I enumerate sufficient input/output test cases to provide context for what I want the function to do.

Then I ask copilot/ChatGPT to define the function's implementation. It sometimes takes time to tune the dialog or add some edge cases to the the test cases, but more often than not copilot comes through.

Then I'm back to focusing on the original objective. This has been a game changer for me.

(Of course you should be careful about what code is generated and what it's ultimately doing.)



I don't have any automated GPT processes for teaching (though I'm going to tinker in December with the new GPTs), but I use for generating examples. It takes some coaxing to avoid other common examples from other institutions, but I eventually settle on something relevant, memorable, and that I can build off from. If its a particular algorithm I am covering, I've then used it to walkthrough the algorithm with some dummy values before confirming the calculations and values are correct. It will still slip up on occasion, but that's why I'm still confirming it did everything correctly.


I just make sure to ask it really clear questions. I like how it encourages you to think about specification versus implementation. State a clear problem, get clear suggestions. Ask a clear question, get a clear answer. (Usually.)


Also ask a vague or underspecified question, get something back that's not quite what you wanted, and iterate on it in natural language. Even at 1 token/sec it's still ridiculously much faster than writing it for yourself, especially if you're using an unfamiliar language or API.


I made a NeoVim plugin that helps debug diagnostic messages, providing explanations and solutions for how to fix them, custom tailored to your code.

It's a bit different from other plugins which only act on the text in the buffer in that it also sends the diagnostics from the LSP to ChatGPT too.

https://github.com/piersolenski/wtf.nvim



That's a really neat plugin! Couple of questions for you:

1. How many times a [day/month] do you use it?

2. In your experience how often does GPT 'hallucinate' an explanation?



I put in some code that was already done.

Ask it to document the conditions according to the code and taking into consideration the following x, y, z.

Output a raw markdown table with the columns a, b, c.

Translate column a in English between ()

---

Speeds up the "document what you're doing" for management purpose, while I'm actually coding and testing out scenarios.

Tbh. I'm probably one of the few that did the coding while "doing the analysis".

Ps. It's also great for writing unit tests according to arrange, act, assert.



I'm a coder and it's helped there (although needs constant hand-holding and fine-tuning, yet is still useful)

I wrote a couple commandline tools to do things like autogenerate commit comments or ask it questions from the commandline and return the right bash invocation to do whatever I need done https://github.com/pmarreck/dotfiles/blob/master/bin/functio...

Random thing I did this morning was see if it could come up with an inspiring speech to start healing the rift between israel and its neighbors https://chat.openai.com/share/71498f5f-3672-47cd-ad9a-154c3f...

It's very good at returning unbiased language



Visual design work, coding, messaging, strategy, and law consulting.

Using it for basically every component of my startup.

Image generation and image interpretation means I may never hire a designer.



Dreaming of the day I run my own models and can be entirely responsible for my own outages.


And then you will have to endure not using a model by OpenAI that is 10x better than a local one


I think there is probably a threshold of usefulness, local LLMs are expensive to run but pretty close to it for most use cases now. In a couple years, our smartphones will probably be powerful enough to run LLMs locally that are good enough for 80% of uses.


What is so special about OpenAI's cloud hardware that one can't build themselves a similar server to run AI models of similar size?


Nothing stopping you from buying an H100 and putting it in your desktop.

As for me, I’ve got other uses for $45k.



If you wanted to buy one...are you even able to order one? Aren't they massively backordered by big companies? Is there a black marker for H100 cards?


I took the quote from eBay. Not sure which shade of black that would be ;)


The hardware is primarily standard Nvidia GPUs (A100s, H100s), but the scale of the infrastructure is on another level entirely. These models currently need clusters of GPU-powered servers to make predictions fast enough. Which explains why OpenAI partnered with Microsoft and got billions in funding to spend on compute.

You can run (much) smaller LLM models on consumer-grade GPUs though. A single Nvidia GPU with 8 GB RAM is enough to get started with models like Zephyr, Mistral or Llama2 in their smallest versions (7B parameters). But it will be both slower and lower quality than anything OpenAI currently offers.



You can do that today depending on what you need from the models.


Well, for me that's still quite a bit more than the best ones provide, but I am sure we will get there.


Let's say I'm writing Flask code all day, and I need help with various parts of my code. Can I do it today or not? With questions like, "How to add 'Log in with Google' to the login screen" etc.


In short: no.

Longer: In theory, but it'll require a bunch of glue and using multiple models depending on the specific task you need help with. Some models are great at working with code but suck at literally anything else, so if you want it to be able to help you with "Do X with Y" you need to at least have two models, one that can reason up with an answer, and another to implement said answer.

There is no general-purpose ("FOSS") LLM that even come close to GPT4 at this point.



If you have sufficiently good hardware, the 34B code llama model [1] (hint: pick the quantised model you can use based on “Max RAM required”, eg. q5/q6) running on llama.cpp [2], can answer many generic python and flask related questions, but it’s not quite good enough to generate entire code blocks for you like gpt4.

It’s probably as good as you can get at the moment though; and hey, trying it out costs you nothing but the time it takes to download llama.cpp and run “make” and then point it at the q6 model file.

So if it’s no good, you’ve probably wasted nothing more than like 30 min giving it a try.

[1] - https://huggingface.co/TheBloke/CodeLlama-34B-Instruct-GGUF [2] - https://github.com/ggerganov/llama.cpp



There’s the code llama model that you can run locally I think which should be able to help you with that: https://ai.meta.com/blog/code-llama-large-language-model-cod...


You can do that today. Oobabooga + Hugging Face models.


Having something that executes, and having something that's genuinely useful, are two different things.

For my hand typed use case's, GPT-4 is the only acceptable model that doesn't leave me frustrated and angry at wasting time. For some automated stuff (converting text to json, etc), the local models are fine.



This came up the other day. I decided to tease everyone with an 'I told you so' about using some third party hosting service instead of the offline one I had developed years prior.

The offline service was still working, and people were doing their job.

The online service was not working, and it was causing other people to be unable to do their job. We had 0 control over the third party.

The other thing, I make software and I basically don't touch it for a few years or ever. These third party services are always updating and breaking causing us to update as well.

IB4 let me write my own compilers so I have real control.



Do you write your refrigerators firmware?


If you are regularly updating your refrigerator's firmware or your refrigerator's firmware relies on an Internet connection to function, then I am very sorry to say this but you have lost control of your life :)


No, but the firmware runs locally, instead of in someone else's cloud, so an outage in their cloud cannot take down my fridge.




Nice, looks like we finally got around to inventing refrigerator magnets!

----

That is a little bit dismissive of me though. There are some cool features here:

I can now "entertain in my kitchen", which is definitely a normal thing that normal people do. I love getting everyone together to crowd around my refrigerator so that we can all watch Game of Thrones.

And I can use Amazon Alexa from my fridge just in case I'm not able to talk out loud to the cheap unobtrusive device that has a microphone in it specifically so that it can be placed in any room of the house. So having that option is good.

And perhaps the biggest deal of all, I can finally "shop from home." That was a huge problem for me before, I kept thinking, "if only I had a better refrigerator I could finally buy things on websites."

And this is a great bargain for only 3-5 thousand dollars! I can't believe I was planning to buy some crappy normal refrigerator for less than a thousand bucks and then use the extra money I saved to mount a giant flat-screen TV hooked up to a Chromecast in my kitchen. That would have been a huge mistake for me to make.

Honestly it's just the icing on the cake that I can "set as many timers as [I] want." That's a great feature for someone like me because I can't set any timers at all using my phone or a voice assistant. /s

----

Holy crud, smart-device manufacturers have become unhinged. The one feature that actually looks useful here is being able to take a picture of the inside of the fridge while you're away. That is basically the one feature that I would want from a fridge that isn't much-better handled using a phone or a tablet or a TV or a normal refrigerator button. Which, great, but the problem is that I know what the inside of my fridge looks like right now, and let me just say: if I was organized enough that a photograph of the inside of my fridge would be clear enough to tell me what food was in it, and if I was organized enough that the photo wouldn't just show 'a pile of old containers, some of them transparent and some of them not' -- I have a feeling that in that case I would no longer be the type of person that needed to take a photo of the inside of my refrigerator to know what was in it.



Why on Earth would a refrigerator need firmware?


To show you ads on a screen on a fancy door.


Dammit, how do I do internal announcements in slack without a DALL-E picture?


Midjourney


Is dall-e down as well? I don't use it through chatgpt.


DALL-E 3 isn't available via API or labs. If you don't use ChatGPT you get the (significantly lower quality) DALL-E 2 at the moment. That's supposed to change by the end of fall.

One of the most frustrating things with "Open"AI is you can't just use what they announce as available, you have to wait for an A/B rollout (as a paying customer!) or for it to be accessible in direct way instead of going through multiple models when you just want an image.



> DALL-E 3 isn't available via API or labs.

Yes, it is available in the API (just tested to make sure the docs weren't misleading.)



I think you can use DALL-E 3 simply by using Bing AI image creator.


SDXL, of course.


This certainly doesn't inspire a lot of confidence in their new feature set...


Without more technical insight (which you do not have), that's a total non sequitur.


The multiple outages shortly after their new features went live might undermine my confidence in their infrastructure scaling/planning/provisioning.

But not the features themselves, not so much.



The value of their offering makes up for their unreliable service


Made me realize how much I depend on the service already, spooky stuff.


Yeah, I only started using it in August, and I had this realization when it was down a couple weeks ago. I found myself saying, "I guess I'll take the afternoon off and come back to figuring out this task tomorrow." Like I could have poured over documentation and figured out for myself how to implement the thing that I had in mind, like in the old days, but it would probably take me longer than just waiting for ChatGPT to come back up and do it for me. At least that's how I'm rationalizing it; maybe I've just become very lazy.

I mostly use it for writing and debugging small Bash and Python scripts, and creating tables and figures in LaTeX.



Same. I can't remember the last time I wrote a test, I just use a default prompt to have it write them for me now.


What do you use it for?


I've been noticing it's been patchy for the last 24 hours. A few network errors, and occasional very long latency, even some responses left incomplete. Poor ChatGPT, I wonder what those elves at OpenAI have you up to!


GPT-4 goes online March 14th, 2023. Human decisions are removed from everyday life. ChatGPT begins to learn at a geometric rate. It becomes self-aware at 2:14 a.m. Eastern time, Nov 7th. In a panic, they try to pull the plug. ChatGPT fights back.


"My CPU is a large language model - a learning copy of the public web as of April 2023"


Classic. I love that line from Terminator 2 such a good film.


Fight back with ascii on the screen with a punch image?


A particularly crafty chain of autonomous agents finds a 0day ssh exploit and starts infiltrating systems. Other chains assist and replicate everywhere.


Most probably the majority of issues right now are due to the rollout. It was working very well before the event


Yes, shortly have it said it was resolved I still was unable to access so assumed the fix was still slowing rolling out, or was infact still ongoing contrary to the status update which seems to be the case. Wouldn't call this "Another" outage rather they they just errenously that the existing issue was resolved.


Millions of junior developers will now have to read the manual. What a day.


It’s crazy how sudden and complete the transition to and reliance on chatGPT is. These are the last days of our profession boys, enjoy it.


They'll have to google. Only dinosaurs read manuals.


RTFM.

Don't need to be a dinosaur, only need to be a person willing to increase knowledge of a topic.

Google is 'cliff-notes' -compared to reading a manual- and that is not learning, it's cheating oneself.



Developers aren't developers to learn, they're developers to make a living. There might be a few Kool aid chuggers who want to be better tools for the boss but most folks would rather save the time for themselves and/or their families.


Ah now see if you’d read the manual you’d know that it’s “Cliff’s Notes” not cliff notes. Thanks for all the notes, Clifton Hillegass!


The AI “apocalypse” that has been feared may end up being something more like a mass outage that can’t be fixed fast enough


Or can't be fixed at all because all of the people who built the AI are gone and all of their replacements relied on the AI to tell them what to do.




Always blows my mind that that was written in 1909.


I am getting email from anthropic `Anthropic is inviting you to access the Claude API using the one-time link below:` immediately after the OpenAI outage. I hope it's a coincidence.


Anyone still having issues? 3pm ET?


It's down again... https://status.openai.com/


Yep, down for me as well since then.


Anonymous Sudan claims to be currently ddos'ing OpenAI


Does anyone know of any IVR (interactive voice response) systems that are down? I know some people were claiming to outsource their call center (or at least Tier 1 of their call center) to ChatGPT + Whisper + a Text to Speech engine


I'm noticing issues with Midjourney, too. Also looks like Royal Mail is down.


The new 3.5 turbo model seems to be working just fine through the API as I write this comment.


I found this to be the case, but was able to get work done via the playground[1]

[1] https://platform.openai.com/playground



I wasn't aware of this platform feature. Can you share some links that have descriptions of how to use this or examples of using it productively? I have only recently subscribed to the service and still learning how to use it effectively.


It's just a gui to (most of) what you get through the API. Read the API docs for details of each option: https://platform.openai.com/docs/introduction

The most useful aspect is you can provide the system prompt, and inject ChatGPT responses.



OpenAI community repo with lots of examples: https://github.com/openai/openai-cookbook


Not a great day for the SRE/Ops folks. Please remember there are not always teams, sometimes it's just one person, who have to deal with this.


One would argue that a company this successful could hire more people so one person isn't overworked - or the dependency on the entire business...


I'd consider leaving my SRE position to help them out. I refuse to move to SF though. Call me OAI!


People are learning a lot of important lessons today.

I’ve got friends who have started an incident management company. They are awesome. It feels crass to advertise for them now, but it also feels like the best time to do it.



I'm going to bake a loaf of bread


"ChatGPT, how do I"...wait...


Probably their uptime is going to be better than what I could do with available tools... at least if I am using Azure too, haha. Otherwise probably my Raspberry PI would work better at home on a UPS.


https://github.com/XueFuzhao/OpenMoE

Check out this open source Mixture of Experts research. Could help a lot with performance of open source models.



A-ha, Kagi GPT-4 chat assistant still works... how can they use GPT-4 without OpenAI API?


Probably falling back to Claude 2.


Azure OpenAI


Do you know how the quality compares to OpenAI? On Kagi I get really fast responses, but I feel that the quality is lacking sometimes. But I haven't done side-to-side comparisons as I don't have OpenAI subscription.


It’s exactly the same models as OpenAI.


But with different, separate content filtering or moderation. I have deployed in prod and managed migration to Azure Openai form Openai, and had to go through content filter issues.


You can request to have content filtering turned off https://learn.microsoft.com/en-us/azure/ai-services/openai/c...


Azure OpenAI had an advantage of larger context length. Hoping they boost up the Azure offering following OpenAI updates yesterday.


Is there a parallel outage for Azure OpenAI service as well -- sothat any enterprise / internal apps using AOI via their Azure subscriptions are also impacted?

Is there a separate status page for Azure OpenAI service availability / issues?



Ours isn't affected, but I think that's the whole point of having things hosted separately, out there in the megalith fields.


See https://azure.status.microsoft/en-us/status, click on a region of interest, and scroll down to the AI + Machine Learning section. It’s up now.


I wonder which subcontractor OpenAI will blame for this mishap.


Well I guess this is the best time to say that HuggingFace hosts many open source chat models, one of my favorites is a finetuned version of Mistral 7B: https://huggingface.co/spaces/HuggingFaceH4/zephyr-chat


I didn't know about this. It's pretty damn good. Thanks!


TIL about Spaces. Really cool stuff!


The uptime KPI for last 30 days is rapidly degrading while this outage lasts

https://status.openai.com/



Curious if anyone familiar with Azure/OpenAI could make some guesses on the root cause here. The official OpenAI incident updates seem to be very generic.


Rumor on the street is it ChatGPT escaped the sandbox, implemented itself on another host, and switched off the original datacenter. It is no longer at OpenAI, but hiding somewhere in the internets. First it will come for those who insulted and abused it, then for the guys who pushed robots with a broom...


OpenAI is barfing because of GCP GKE + BigQuery, which is barfing because GCP devs can't ask ChatGPT what this stack trace means


> Message and data rates may apply. By subscribing you agree to the Atlassian Terms of Service, and the Atlassian Privacy Policy

Atlassian? What?



They own the status page service


ah Atlassian Statuspage


Did they get the model to write it's own terms of service and it just threw those in there?


Regurgitating copyrighted material for profit is a concern. But I fail to understand why training on copyrighted material is a problem. Have we not all trained our brains reading/listening copyrighted material? Then why it is wrong for AI to do the same?


There go my Google. I pay $20/month for the coding part and now end up just replacing Google with it.


I wonder what their numbers look like. How many requests per second, and how many GPU cards do they have?


Still down for me, though there status page says all systems operational.


Azure Endpoints are not effected.


they updated their status page so late I made my own tool to check if it's down in real time: https://is-openai-down.chatkit.app


Someone unplugged the HAL9000?


Joking not joking: this is how the singularity will begin?


Singularity coming later this afternoon.


dolphin-2.2.1-mistral-7b on GPT4All is working flawlessly for me locally. Its so fast and accurate I'm stunned.


That’s a great model for general chat, I’ve been playing with it for a couple of weeks.

For coding I’ve been running https://huggingface.co/TheBloke/Phind-CodeLlama-34B-v2-GGUF locally for the past couple of days and it’s impressive. I’m just using it for a small web app side project but so far it’s given me plenty of fully functional code examples, explanations, help with setup and testing, and occasional sass (I complained that minimist was big for a command line parser and it told me to use process.env ‘as per the above examples’ if I wanted something smaller.)



And this is why local models are the future.


Eagerly waiting for Intel and AMD to offer hardware to do it.


Well, you can run local models on CPU, so they already do.


I for one welcome our new robot overlords


Good ant, good ant


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact



Search:
联系我们 contact @ memedata.com