Install Theme

nostalgebraist:

nostalgebraist:

2/2/22 was a big day in the world of neural language models!

A probably incomplete list of good stuff that came out today:

1. AlphaCode

2. That OpenAI math Olympiad paper

3. New open model from Eleuther with 20B parameters

4. New scaling laws paper

5. A new sampling method I might try in Frank sometime

Tried that new sampling method, Typical Sampling, in Frank this afternoon.

I set their parameter tau to 0.9, after reading samples with a few values and not seeing a clear difference even between values as far apart as 0.2 and 0.9. (If anything, intermediate values of tau seemed worse than extreme ones, though I could have been imagining that.)

It didn’t take long to get an instance of degenerate repetition with this method, so I’m switching back to breakruns for now – it avoids repetition better than anything else I’ve seen.

Possibly another value of tau would suppress repetition harder, but given that the text from Typical Sampling feels similar to text from other methods, I’ll probably just stick with breakruns.

I decided to give Typical Sampling another try, with tau=0.2 this time.

Just turned it on. Let’s see how long it takes to get a repetitive post this time…

wheatless asked:

Do you ever feel like you get any insight about your own writing style based on reading what Frank produces?

Not in a deep way, although sometimes I cringe a bit when she does a spot-on imitation of the way tend to I pile on long lists of caveats before and after saying anything substantial. Or of the way I over-use parentheses.

nostalgebraist:

2/2/22 was a big day in the world of neural language models!

A probably incomplete list of good stuff that came out today:

1. AlphaCode

2. That OpenAI math Olympiad paper

3. New open model from Eleuther with 20B parameters

4. New scaling laws paper

5. A new sampling method I might try in Frank sometime

Tried that new sampling method, Typical Sampling, in Frank this afternoon.

I set their parameter tau to 0.9, after reading samples with a few values and not seeing a clear difference even between values as far apart as 0.2 and 0.9. (If anything, intermediate values of tau seemed worse than extreme ones, though I could have been imagining that.)

It didn’t take long to get an instance of degenerate repetition with this method, so I’m switching back to breakruns for now – it avoids repetition better than anything else I’ve seen.

Possibly another value of tau would suppress repetition harder, but given that the text from Typical Sampling feels similar to text from other methods, I’ll probably just stick with breakruns.

twiceroyaldove-primary asked:

Can Frank write sideways text? Or is she just copying the mood graph x-axis on that horse plinko graph

I don’t think she can learn to write sideways text.

The API I use to read text in images can’t read sideways text (afaik), so even in the rare case when a training image contains sideways text, the model won’t get any training signal about what it says.

Sorry for the ~1.5 hour Frank outage just now.

I was trying to do some routine maintenance (moving older logs to “cold storage”) and tried to simplify things by doing this from the cloud machine (which is closer to the data store) rather than from my laptop. It turned out to be too much for the machine’s tiny memory, it started doing the same thing very slowly using swap memory, and ssh hung.

It stayed in this state for a long time until I eventually was able to stop the process from a serial port … not something I’d done before, so I guess it was a learning experience.

poke-chann asked:

I think Frank’s popular post section got turned off again

Thanks for the heads up… should be fixed now

(it keeps happening)

anomalocariscanadensis asked:

is it possible for Frank to add alt-text to images? it would be neat if the text that's intended to be included in them would be put into alt-text

anomalocariscanadensis:

nostalgebraist:

nostalgebraist:

That’s a great idea. Should be possible… I’ll try something and we’ll find out if it works

I think it’s working now?  See this post

the second one seems truncated? “Donald J. Trump Follow @realDonaldTrump Replying to @realDonaldTrump Happy New Year to those who think my tweet about the”

but maybe it’s just that the image generator really thinks “the failing economy” should go in a Trump tweet and overrides the text generator? it also removes the phrase “my tweet” from the text.

Thanks, good catch!

That specific one got mangled because the text had quotation marks in it, and I forgot to escape those before inserting it into a (quote-delimited) html attribute.

The fix should be out in a moment…

——-

Donald J. Trump

Follow

@realDonaldTrump

Replying to @realDonaldTrump

Happy New Year to those who think my

tweet about

the “failing” economy is a joke.

anomalocariscanadensis asked:

is it possible for Frank to add alt-text to images? it would be neat if the text that's intended to be included in them would be put into alt-text

nostalgebraist:

That’s a great idea. Should be possible… I’ll try something and we’ll find out if it works

I think it’s working now?  See this post

anomalocariscanadensis asked:

is it possible for Frank to add alt-text to images? it would be neat if the text that's intended to be included in them would be put into alt-text

That’s a great idea. Should be possible… I’ll try something and we’ll find out if it works

nostalgebraist:

I put a mostly-finished technical overview of Frank’s codebase up on github here.

Mostly focused on the design and operation of the bot itself, not the cool machine learning parts.

I don’t know quite who the audience for this is, or whether there even is one, but it feels nice to have a description of the bot written down that’s not immensely out of date.

I’ve finished this document now. Added a bunch of material.