Yazılar

ChatGPT Introduces Library Feature for Easy Access to AI-Generated Images

OpenAI Launches New Library Feature to Organize AI-Generated Images in ChatGPT

OpenAI has introduced a new library feature within ChatGPT that provides users with a centralized space to view all their AI-generated images. Announced on Wednesday, the feature is now available across all ChatGPT platforms — web, desktop, and mobile — for registered users. The library is designed to help users easily browse, revisit, and reuse their previously created images without digging through old chat histories. In addition to viewing, the update also offers editing capabilities directly from the library interface.

The feature was officially revealed via OpenAI’s post on X (formerly Twitter), highlighting its broad availability to both free users and those subscribed to the Plus and Pro plans. Accessible via the left-hand sidebar on web and mobile apps, the library displays only images generated using GPT-4o’s image creation capabilities. Images created with earlier models like DALL-E are not included in this view, according to OpenAI’s support documentation.

Inside the library, users will find a new “Make Image” button at the bottom, offering a quick way to jump back into generating fresh visuals. When a user taps and holds on an existing image, it enlarges in a separate window where four new options appear: Edit, Select, Save, and Share. Saving allows users to download the image locally, while sharing integrates with third-party apps to send images to friends and social media.

The editing tools add even more flexibility. Selecting “Edit” opens a new chat where the image is attached, allowing users to apply further text-based prompts for significant modifications or to generate related creations. The “Select” tool provides more granular control, letting users highlight and modify specific parts of an image. An adjustable slider refines selection sizes, and Undo/Redo options streamline the editing process. Additionally, a Copy button lets users quickly add images to their clipboard for use elsewhere. Together, these new features mark a major step forward in making image generation within ChatGPT more organized and interactive.

Appy Pie Unveils PixelForge and Vibeo AI Models for Image and Video Creation

Appy Pie, a leading Indian no-code platform specializing in artificial intelligence (AI), has introduced two groundbreaking AI models: PixelForge and Vibeo. These multimodal large language models (LLMs) are designed to revolutionize how images and videos are created. PixelForge, as a text-to-image generation model, enables users to transform text prompts into high-resolution, photorealistic, and artistic visuals. On the other hand, Vibeo takes things a step further by generating videos from text or image inputs, offering even greater versatility in multimedia creation. These models are being made available to both individual users and businesses through Appy Pie’s comprehensive Appy Pie Design platform, which also supports the development of mobile apps, websites, and AI-driven chatbots.

The new models, PixelForge and Vibeo, are the result of Appy Pie’s in-house development, marking a significant departure from their earlier text-focused AI tool, Flawless Text. The company asserts that these two new models are more advanced, catering not just to creators but also to marketing professionals and enterprises that require dynamic and customizable visual content. PixelForge stands out for its ability to generate a wide array of image styles, making it a versatile tool for any project, whether artistic or professional. Meanwhile, Vibeo offers a compelling solution for those looking to create videos with just a simple text or image input.

PixelForge’s core feature is its ability to generate high-quality images from text descriptions. It supports a diverse range of visual styles and can cater to various compositions and use cases, offering something for everyone, from graphic designers to content creators. While the company has highlighted similarities with popular models like OpenAI’s DALL-E and Stability AI’s Stable Diffusion, it has yet to release detailed benchmark data to support these claims. However, Appy Pie promises that PixelForge is optimized for a seamless user experience with a focus on both speed and creativity. Despite the lack of technical details, such as resolution outputs and rate limits, PixelForge is poised to be an invaluable tool in the growing field of AI-powered content creation.

Vibeo, the video generation model, takes AI capabilities a step further by providing users with the ability to generate videos from either textual prompts or reference images. This model is specifically designed to prioritize realism, ensuring that the generated videos not only match the user’s expectations but also convey the intended mood and motion. With Vibeo, users can create dynamic video content with minimal effort, making it an ideal tool for everything from marketing materials to social media content. As Appy Pie continues to innovate in the AI space, these models could reshape the future of multimedia content production, offering users the tools to produce high-quality images and videos with just a few simple inputs.

Anthropic Allegedly Developing Voice Mode Feature for Claude AI

Anthropic is reportedly working on a highly anticipated voice mode feature for its AI chatbot, Claude. The company, based in San Francisco, is expected to launch the new feature as early as this month, marking a significant shift for the AI firm. While competitors like OpenAI and Google have already integrated voice capabilities into their chatbots—such as ChatGPT’s voice feature and Gemini’s similar tool—Claude has so far only offered text-based interactions. This move comes shortly after Anthropic introduced an educational subscription plan, mirroring OpenAI’s Edu offering, signaling the company’s broader push into more dynamic AI tools.

The new voice mode feature is expected to be rolled out gradually, with a Bloomberg report suggesting that the feature could begin rolling out in April. Initially, it will be available to a select group of users, with plans possibly subject to change. The inclusion of voice capabilities would place Claude on a more competitive footing with its peers, allowing users to interact with the AI in a more natural, conversational manner. The voice mode is likely to make the AI experience more immersive, combining the capabilities of voice recognition with Claude’s advanced text-based responses.

According to sources familiar with the development, the feature will include three distinct voices: Airy, Mellow, and Buttery. Notably, Buttery is expected to feature a British accent, adding a unique element to the AI’s vocal range. The discovery of this feature was first noted by an app researcher named “M1Astra,” who found clues about the voices in the code of Claude’s iOS app. However, details about the voice mode remain sparse, and it is unclear whether the feature will serve as a basic text-to-speech function or if it will feature more advanced, human-like voice synthesis, akin to ChatGPT’s more sophisticated voice interaction system.

Anthropic’s delayed entry into the voice chatbot arena comes as major players in the AI space, including OpenAI, Google, and Microsoft, have already rolled out voice-based features. Meta, too, is reportedly developing a two-way voice chat mode for its Meta AI, further intensifying the competition. As Anthropic looks to add this new functionality to Claude, it will be interesting to see how the feature stacks up against the already established voice capabilities of its rivals. The feature’s availability to all users or its potential restriction to premium subscribers is also yet to be determined, leaving room for further speculation about the company’s future plans.