Software developer at a big library, cyclist, photographer, hiker, reader. Email: chris@improbable.org
24193 stories
·
216 followers

National Archives Bans Employee Use of ChatGPT

1 Share

The National Archives and Records Administration (NARA) told employees Wednesday that it is blocking access to ChatGPT on agency-issued laptops to “protect our data from security threats associated with use of ChatGPT,” 404 Media has learned.

“NARA will block access to commercial ChatGPT on NARANet [an internal network] and on NARA issued laptops, tablets, desktop computers, and mobile phones beginning May 6, 2024,” an email sent to all employees, and seen by 404 Media, reads. “NARA is taking this action to protect our data from security threats associated with use of ChatGPT.”

The move is particularly notable considering that this directive is coming from, well, the National Archives, whose job is to keep an accurate historical record. The email explaining the ban says the agency is particularly concerned with internal government data being incorporated into ChatGPT and leaking through its services. “ChatGPT, in particular, actively incorporates information that is input by its users in other responses, with no limitations. Like other federal agencies, NARA has determined that ChatGPT’s unrestricted approach to reusing input data poses an unacceptable risk to NARA data security,” the email reads.

The NARA email also references previous guidance given to employees of NARA about the “responsible use of artificial intelligence” which includes instructions like “Do not rely on LLMs for factual information” and explains that LLMs “should always be used with caution.”

“NARA information should never be used with chatbots or other online AI applications. ChatGPT and other similar tools may produce results with biases, including racial and gender biases, in their responses. They may generate false or misleading responses that may negatively impact the quality of work performed or assisted by AI tools,” previous guidance, which 404 Media obtained from NARA (and is available as a PDF below), reads. “Importantly, AI-enabled tools incorporate the inputs and responses of previous user queries into their responses to queries from other users. This means that information provided by NARA users will be incorporated into other responses, which may provide unrelated users with biased and otherwise inaccurate responses about NARA or using NARA information.”

The email goes on to explain that “If sensitive, non-public NARA data is entered into ChatGPT, our data will become part of the living data set without the ability to have it removed or purged.” NARA said in the email that it is “exploring the use of other AI solutions, such as Microsoft Copilot and Google Gemini, which provide service similar to ChatGPT, but in a more controlled environment. These tools differ from ChatGPT because they protect data input by federal agencies placing it in a private repository that is not shared with others.”

Last year, the Biden administration directed federal agencies to “ensure the U.S. government is leading by example on mitigating AI risks and harnessing AI opportunities” by studying AI and creating policies for its government use. Other federal government agencies, including the Department of Energy, the Department of Veterans Affairs, the Department of Agriculture (USDA), and the Social Security Administration have also blocked access to ChatGPT for its employees, with each of them citing data privacy concerns. The USDA’s guidance specifically noted that “While Generative AI models and tools show promise, there are some concerning characteristics, such as generating misinformation, hallucinations, inaccurate or outdated responses, lack of data privacy protections, and potential misuse.” 

The US Government Accountability Office also specifically warned that “these systems can also generate ‘hallucinations’—misinformation that seems credible—and can be used to purposefully create false information.”

Update: This article has been updated to include previous NARA guidance on AI obtained by 404 Media today.



Read the whole story
Share this story
Delete

Printing music with CSS grid

2 Shares

Too often have I witnessed the improvising musician sweaty-handedly attempting to pinch-zoom an A4 pdf on a tiny mobile screen at the climax of a gig. We need fluid and responsive music rendering for the web!

Stephen Band

Read the whole story
Share this story
Delete

Jane Street is big. Like, really, really big

1 Share
Read the whole story
Share this story
Delete

Utah woman charged with sexual battery after pulling teen's skirt down

1 Share

A 48-year-old Utah woman featured in a viral TikTok video accusing her of being a “Karen” was charged with sexual battery this week after she told police she pulled a teen’s skirt down at a St. George steakhouse because she felt it was “inappropriate.”

The woman was arrested days after she called police herself, reporting last Saturday that the TikTok video amounted to a “threat on her life” because it insinuated that her interaction with the 19-year-old at least two hours earlier was “sexual in nature,” police documents state.

The video posted to TikTok did not show the woman pulling down the teen’s skirt, but it did show her interacting with the teen’s group in the lobby of Sakura Japanese Steakhouse after the alleged assault.

“I happen to work for the state, and if I have to watch your a-- cheeks hanging out again, I will call CPS,” the 48-year-old woman can be heard saying in the video, seeming to refer to the Utah Division of Child and Family Services.

“She’s over 18,” one of the people in the teen’s group responded.

“She is 19 years old,” another said.

“You don’t get to touch her,” someone added.

It’s unclear whether the 48-year-old woman works for the state of Utah, and if so, in what capacity. The Salt Lake Tribune generally does not name defendants unless they have been charged with a felony. The sexual battery charge the woman faces is a class A misdemeanor.

The woman’s name was not found within an initial search of public salary data. A state employee directory, however, lists an email and phone number for a person with the same name and indicates the individual works with the Utah attorney general’s office.

More details about the person’s role are not listed. The woman’s name was not found in a Utah State Bar directory search. When reached for comment Friday, a spokesperson with the Utah attorney general’s office did not provide more information.

“We are aware of the situation and are following [attorney general’s office] policies and procedures,” the spokesperson said in a statement.

The woman’s 911 call came about two hours after she had initially called police about the teen, reporting that the teen’s skirt was hiked up above her genitals while in the view of several minors, police documents state. The woman felt steakhouse staff and other adults were not addressing the issue.

In that call, she told police she “pulled [the teen’s] skirt down and told her to be aware of what she was showing,” according to a probable cause statement for the woman’s arrest.

No officers went to the restaurant after her first call, police records indicate, and no officers went to meet her after her 911 call about the video.

Instead, a St. George police officer made contact with the woman the next day. The woman told the officer she felt it was “her responsibility” to approach the teen and pull her skirt down, because a young boy had seemed to point at the teen’s skirt, and the boy’s father — neither of whom the 48-year-old had any relation to — “did nothing about it,” the document states.

She explained that her intent behind threatening to call “CPS” was to file a report alleging the teen had indecently exposed herself to the boy.

The woman added that she believed the 19-year-old she approached was a minor. The officer replied that “the belief the victim was a minor, should have been more reason to not touch the victim,” the document states.

The officer also countered that while the woman asserted she could see the teen was nude under her skirt, the teen provided evidence to suggest she was wearing underwear and shorts under her skirt and was not being lewd in public.

The woman also argued she never touched the teen, “she had only touched the female’s skirt,” the officer wrote.

“I explained to [the woman] that she had still engaged in criminal behavior by touching the female’s clothing, and that her behavior was not appropriate,” the officer continued.

But he did not arrest her that day. The next day, on Monday, the young woman whose skirt was pulled down came forward, reporting to investigators that the 48-year-old woman had sexually assaulted her by putting her hands up her skirt and groping her, police documents state.

Seven other witnesses came forward and filed statements consistent with the young woman’s report. The teen told authorities that her friend had posted the TikTok video in an attempt to identify the woman who touched her. The teen also provided police with a separate video of herself in the skirt she wore that night.

The 48-year-old woman was arrested Wednesday and charged with sexual battery on Thursday. She was booked into the Washington County jail after her arrest, but released on bail Thursday after paying $1,000.

— Tribune staff writer Emily Anderson Stern contributed to this report.

Read the whole story
Share this story
Delete

Actually Using SORA - fxguide

1 Share

In February, we pushed our first story on SORA; OpenAI had just released the first clips from SORA, which we described at the time as the video equivalent of DALL·E for videos. SORA is a diffusion model that generates videos significantly longer and with more cohesion than any of its rivals. By giving the model foresight of many frames at a time, they have solved the challenging problem of ensuring a subject stays consistent even when it goes out of view temporarily. SORA can generate entire videos, all at once up to a minute in length. At the time, OpenAI also published technical notes indicating that it could (in the future) extend generated videos to make them longer or blend two videos seamlessly.

Several select production teams have been given limited access to SORA in the last few weeks. One of the most high-profile was the Shy Kids team, who produced the SORA short film Air Head. Sidney Leeder produced the film.  Walter Woodman was the writer and director, while Patrick Cederberg was responsible for the post-production. The Toronto team have been nicknamed “punk-rock Pixar”, while their work has garnered Emmy nominations and been long-listed for the Oscars. We sat down this week with Patrick for a long chat about the current state of SORA.

Shy Kids is a Canadian production company renowned for its eclectic and innovative approach to media production. Originating as a collective of creatives from various disciplines, including film, music, and television, Shy Kids has gained recognition for its unique narrative styles and engaging content. The company often explores adolescence, social anxiety, and the complexities of modern life while maintaining a distinctively whimsical and heartfelt tone. Their work showcases a keen eye for visual storytelling and often features a strong integration of original music, making their productions resonant and memorable. Shy Kids has successfully carved out a niche by embracing new AI technology and creativity, pushing what is possible.

SORA : Mid-April ’24.

SORA is in development and is actively being improved through feedback from teams such as Shy Kids, but here is how it currently works. It is important to appreciate that SORA is effective almost pre-alpha. It has not been released nor is it in beta.

“Getting to play with it was very interesting,” Patrick comments. “It’s a very, very powerful tool that we’re already dreaming up all the ways it can slot into our existing process. But I think with any generative AI tool; control is still the thing that is the most desirable and also the most elusive at this point.”

UI

The user interface allows an artist to input a text prompt; OpenAI’s ChatGPT then converts this into a longer string, which triggers the clip generation. At the moment, there is no other input; it is yet to be multimodal. This is significant as while SORA is correctly applauded for its object consistency during a shot, but there is nothing to help make anything from the first shot match in a second shot. The results would be different even if you ran the same prompt a second time. “The closest we could get was just being hyper-descriptive in our prompts,” Patrick explains. “Explaining wardrobe for characters, as well as the type of balloon, was our way around consistency because shot to shot / generation to generation, there isn’t the feature set in place yet for full control over consistency.”

The individual clips are remarkable and jaw-dropping for the technology they represent, but the use of the clips depends on your understanding of implicit or explicit shot generation. Suppose you ask SORA for a long tracking shot in a kitchen with a banana on a table. In that case, it will rely on its implicit understanding of ‘banana-ness’ to generate a video showing a banana. Through training data, it has ‘learnt’ the implicit aspects of banana-ness: such as ‘yellow’, ‘bent’, ‘has dark ends’, etc. It has no actual recorded images of bananas. It has no ‘banana stock library’ database; it has a much smaller compressed hidden or ‘latent space’ of what a banana is.  Every time it runs, it shows another interpretation of that latent space. Your prompt replies on an implicit understanding of banana-ness.

Prompting the right thing to make Sonny

For Air Head, the scenes were made by generating multiple clips to an approximate script, but there was no explicit way to have the actual yellow balloon head the same from shot to shot. Sometimes, when the team prompted for a yellow balloon, it wouldn’t even be yellow. Other times, it had a face embedded in it or a face seemingly drawn on the front of the balloon. As many balloons have string, often the Air Head character, nicknamed Sonny, the balloon guy, would have a string down the front of the character’s shirt. Since it implicitly links string with balloons and thus these would need to be removed in post.

Resolution

Air Head is only using SORA-generated footage, but much of it was graded, treated, and stabilised, and all of it was upscaled or upresed. The clips the team worked with were generated at a lower resolution and then upresed using AI tools outside SORA or OpenAI. “You can do up to 720 P (resolution),” Patrick explains. “I believe there’s a 1080 feature that’s out, but it takes a while (to render). We did all of Air Head at 480 for speed and then upright using Topaz.”

Prompting ‘time’: A slot machine.

The original prompt is automatically expanded but also displayed along a timeline. “You can go into those larger keyframes and start adjusting information based on changes you want generated.” Parick explains, “There’s a little bit of temporal control about where these different actions happen in the actual generation, but it’s not precise… it’s kind of a shot in the dark – like a slot machine – as to whether or not it actually accomplishes those things at this point.” Of course, Shy Kids were working with the earliest of prototypes, and SORA is still constantly being worked on.

In addition to choosing a resolution, SORA allows the user to pick the aspect ratio, such as portrait or landscape (or square). This came in handy on the shot that pans up from Sonny’s jeans to his balloon head. Unfortunately, SORA would not render such a move natively, always wanting the main focus of the shot—the balloon head—to be in the shot. So the team rendered the shot in portrait mode and then manually, via cropping, created the pan-up in post.

Prompting camera directions

For many genAI tools, a valuable source of information is the metadata that comes with the training data, such as camera metadata. For example, if you train on still photos, the camera metadata will provide the lens size, the f-stop and many other critical pieces of information for the model to train on. With cinematic shots, the ideas of ‘tracking’, ‘panning’, ’tilting’ or ‘pushing in’ are all not terms or concepts captured by metadata. As much as object permanency is critical for shot production, so is being able to describe a shot, which Patrick noted was not initially in SORA. “Nine different people will have nine different ideas of how to describe a shot on a film set. And the (OpenAI) researchers, before they approached artists to play with the tool, hadn’t really been thinking like filmmakers.” Shy Kids knew that their access was very early, but “the initial version about camera angles was kind of random.” Whether or not SORA was actually going to register the prompt request or understand it was unknown as the researchers had just been focused on image generation. Shy Kids were almost shocked by how much the OpenAI was surprised by this request. “But I guess when you’re in the silo of just being researchers, and not thinking about how storytellers are going to use it… SORA is improving, but I would still say the control is not quite there. You can put in a  ‘Camera Pan’ and I think you’d get it six out of 10 times.”  This is not a unique problem nearly all the major video genAI companies are facing the same issue. Runway AI is perhaps the most advanced in providing a UI for describing the camera’s motion, but Runway’s quality and length of rendered clips are inferior to SORA.

Render times

Clips can be rendered in varying segments of time, such as 3 secs, 5 sec, 10 sec, 20sec, up to a minute. Render times vary depending on the time of day and the demand for cloud usage. “Generally, you’re looking at about 10 to 20 minutes per render,” Patrick recalls. “From my experience, the duration that I choose to render has a small effect on the render time. If it’s 3 to 20 seconds, the render time tends not to vary too much from between a 10 to 20-minute range. We would generally do that because if you get the full 20 seconds, you hope you have more opportunities to slice/edit stuff out and increase your chances of getting something that looks good.”

Roto

While all the imagery was generated in SORA, the balloon still required a lot of post-work. In addition to isolating the balloon so it could be re-coloured, it would sometimes have a face on Sonny, as if his face was drawn on with a marker, and this would be removed in AfterEffects. similar other artifacts were often removed.

Editing a 300:1 shooting ratio

The Shy Kids methodology was to approach post-production and editing like a documentary, where there is a lot of footage, and you weave a story from that material rather than strictly shooting to a script. There was a script for the short film, but the team needed to be agile and adapt. “It was just getting a whole bunch of shots and trying to cut it up in an interesting way to the VO,” Patrick recalls.

For the minute and a half of footage that ended up in the film, Patrick estimated that they generated “hundreds of generations at 10 to 20 seconds a piece”. Adding, “My math is bad, but I would guess probably 300:1 in terms of the amount of source material to what ended up in the final.”

Comping multiple takes and retiming

On Air Head, the team did not comp multiple takes together. For example, the shots of the balloon drifting over the motor racing were all generated in the one shot fairly much as seen. However, they are working on a new film that mixes and composites multiple takes into one clip.

Interestingly, many of the Air Head clips were generated as if shot in slow motion, while this was not requested in the prompt. This happened for unknown reasons, and so many of the clips had to be retimed to appear to have been shot in real-time. Clearly, this is easier to do than the reverse of slowing down rapid motion, but still, it seems like an odd aspect to have been inferred from the training data. “I don’t know why, but it does seem like a lot of clips at 50 to 75% speed,” he adds. “So there was quite a bit of adjusting timing to keep it all from feeling like a big slowmo project.”

Lighting and grading

Shy Kids used the term ‘35 mm film‘ in their prompts as a keyword and generally found that the prompt 35mm gave a level of consistency that they sought. “If we needed a high contrast, we could say high contrast, and say key lighting would generally give us something that was close,” says Patrick. “We still had to take it through a full color grade, and we did our own digital filmic look, where we applied grain and flicker to just sort of meld it all together.” There is no option for additional passes such as mattes or depth passes.

Copyright

OpenAI is trying to be respectful and not allow material to be generated that violates copyright or produces images that would appear to be from someone they are not. For example, if you prompt something such as 35mm film in a futuristic spaceship, a man walks forward with a light sword,  SORA will not allow the clip to be generated as it is too close to Star Wars. But the Shy Kids accidentally bumped into this during early testing. Patrick recalls that when they initially sat down and just wanted to test SORA, “We had that one shot behind the character’s back; it’s kind of that Aronofsky following shot. And I think it was just my dumb brain, as I was tired, but I put ‘Aronofsky type shot’ in and got hit with a can’t do that.,” he recalls. Hitchcock Zoom was another thing that came up as something that is  now by osmosis, a technical term, but SORA would reject the prompt for copyright purposes.

Sound

Shy Kids are known for their audio skills in addition to their visual skills. The music in the short film is their own. “It was a song we had in the back catalogue that we almost immediately decided on because the song’s called The Wind, ” says Patrick. “We all just liked it.”

Patrick himself is the voice of Sonny. “Sometimes we’d feel pacing-wise the film needed another beat. So I would write another line, record it, and come up with some more SORA generations, which is another powerful use of the tool in the post: when you’re in a corner, and you need to fill a gap, it’s a great way to start brainstorming and just spit clips out to see what you can use to fill the pacing problem.”

Summary

SORA is remarkable; the Shy Kids team produced Air Head with a team of just 3 people in around 1.5 to 2 weeks. The team is already working on a wonderful, self-aware, and perhaps ironic sequel. “The follow-up is a journalistic approach to Sonny, the balloon guy, and his reaction to fame and subsequent sort falling out with the world,” says Patrick. “And we’re exploring new techniques!” The team is looking to be a bit more technical in their experimentation, incorporating AE composting of SORA elements into real live-action footage and using SORA as a supplementary VFX tool.

SORA is very new, and even the basic framework that OpenAI has sketched out and demonstrated for SORA has yet to be available for early tests to use. It is doubtful that SORA in its current form will be released anytime soon, but it is an incredible advance in a particular type of implicit image generation. For high-end projects, it may be a while before it allows the level of specificity that a director requires. It will be more than ‘close enough’ for many others while delivering stunning imagery. Air Head still needed a large amount of editorial and human direction to produce this engaging and funny story film. “I just feel like people have to SORA as an authentic part of their process; however, if they don’t want to engage with anything like that, that’s fine too.”

Read the whole story
Share this story
Delete

We can have a different web

2 Shares
Read the whole story
Share this story
Delete
Next Page of Stories