custom gpt
Why Nicholas Thompson Made a Custom GPT to Run Faster
The Atlantic CEO's new book,, examines his complicated relationship with the sport. On this week's episode of, he talks about the ways tech is helping him become a better runner. To most of the world, Nicholas Thompson is known as an editor, an AI enthusiast, or something of a LinkedIn influencer. But the former WIRED editor in chief, who is now CEO of The Atlantic, is often better known to colleagues as . On Tuesday, Thompson is releasing . As the title suggests, it's a book about his commitment to running--Thompson runs a ridiculously fast marathon and holds the American 50K record for the 45-49 age group. Ultimately, though, the book examines the complicated relationship between the sport, Thompson, and his father, who first took him on a run when he was just 5 years old. Tech obsessives, of course, will also get their fix: includes plenty of science-backed training guidance and documents Thompson's experience training with elite Nike coaches. On this week's episode of, I talked to Thompson (who was also my first boss; he hired me as an intern at WIRED in 2008) about his book, the interplay between running and addiction, and what he thinks AI can do for runners for writers. It is a joy to be here with you at Condé Nast at WIRED. I loved coming up those elevators. I love seeing you as the editor in chief. I'm thrilled that you're here. We're going to start this conversation the way we start all of them, which is with a little warmup, some rapid-fire questions. In honor of your new book,, I'm gonna make them entirely running themed. I mean, if your listeners don't wanna hear about running Trail run or track run? Worst running injury you've ever had. The one you wish people would stop talking to you about. You only need to run a 20-miler before a marathon. What do you need to run? Why do people die at mile 20? Because they only train for [marathons] with 20-mile-runs. I generally prefer people, but then you have to schedule it. Backup sport of choice if you could never run again.
- North America > United States > New York (0.05)
- South America (0.04)
- North America > United States > Massachusetts > Middlesex County > Medford (0.04)
- (7 more...)
- Leisure & Entertainment > Sports > Running (1.00)
- Health & Medicine > Therapeutic Area (1.00)
- Health & Medicine > Consumer Health (1.00)
- Information Technology > Communications > Social Media (1.00)
- Information Technology > Communications > Mobile (1.00)
- Information Technology > Artificial Intelligence (1.00)
Privacy and Security Threat for OpenAI GPTs
Wenying, Wei, Kaifa, Zhao, Lei, Xue, Ming, Fan
Large language models (LLMs) demonstrate powerful information handling capabilities and are widely integrated into chatbot applications. OpenAI provides a platform for developers to construct custom GPTs, extending ChatGPT's functions and integrating external services. Since its release in November 2023, over 3 million custom GPTs have been created. However, such a vast ecosystem also conceals security and privacy threats. For developers, instruction leaking attacks threaten the intellectual property of instructions in custom GPTs through carefully crafted adversarial prompts. For users, unwanted data access behavior by custom GPTs or integrated third-party services raises significant privacy concerns. To systematically evaluate the scope of threats in real-world LLM applications, we develop three phases instruction leaking attacks target GPTs with different defense level. Our widespread experiments on 10,000 real-world custom GPTs reveal that over 98.8% of GPTs are vulnerable to instruction leaking attacks via one or more adversarial prompts, and half of the remaining GPTs can also be attacked through multiround conversations. We also developed a framework to assess the effectiveness of defensive strategies and identify unwanted behaviors in custom GPTs. Our findings show that 77.5% of custom GPTs with defense strategies are vulnerable to basic instruction leaking attacks. Additionally, we reveal that 738 custom GPTs collect user conversational information, and identified 8 GPTs exhibiting data access behaviors that are unnecessary for their intended functionalities. Our findings raise awareness among GPT developers about the importance of integrating specific defensive strategies in their instructions and highlight users' concerns about data privacy when using LLM-based applications.
- Asia > China > Hong Kong (0.04)
- Asia > China > Shaanxi Province > Xi'an (0.04)
- North America > United States (0.04)
- Asia > China > Guangdong Province > Shenzhen (0.04)
- Information Technology > Security & Privacy (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.63)
A Large-Scale Empirical Analysis of Custom GPTs' Vulnerabilities in the OpenAI Ecosystem
Ogundoyin, Sunday Oyinlola, Ikram, Muhammad, Asghar, Hassan Jameel, Zhao, Benjamin Zi Hao, Kaafar, Dali
Millions of users leverage generative pretrained transformer (GPT)-based language models developed by leading model providers for a wide range of tasks. To support enhanced user interaction and customization, many platforms-such as OpenAI-now enable developers to create and publish tailored model instances, known as custom GPTs, via dedicated repositories or application stores. These custom GPTs empower users to browse and interact with specialized applications designed to meet specific needs. However, as custom GPTs see growing adoption, concerns regarding their security vulnerabilities have intensified. Existing research on these vulnerabilities remains largely theoretical, often lacking empirical, large-scale, and statistically rigorous assessments of associated risks. In this study, we analyze 14,904 custom GPTs to assess their susceptibility to seven exploitable threats, such as roleplay-based attacks, system prompt leakage, phishing content generation, and malicious code synthesis, across various categories and popularity tiers within the OpenAI marketplace. We introduce a multi-metric ranking system to examine the relationship between a custom GPT's popularity and its associated security risks. Our findings reveal that over 95% of custom GPTs lack adequate security protections. The most prevalent vulnerabilities include roleplay-based vulnerabilities (96.51%), system prompt leakage (92.20%), and phishing (91.22%). Furthermore, we demonstrate that OpenAI's foundational models exhibit inherent security weaknesses, which are often inherited or amplified in custom GPTs. These results highlight the urgent need for enhanced security measures and stricter content moderation to ensure the safe deployment of GPT-based applications.
- Oceania > Australia > New South Wales > Sydney (0.04)
- Europe > Switzerland (0.04)
- Africa > Nigeria > Lagos State > Lagos (0.04)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (1.00)
Towards Safer Chatbots: A Framework for Policy Compliance Evaluation of Custom GPTs
Rodriguez, David, Seymour, William, Del Alamo, Jose M., Such, Jose
Large Language Models (LLMs) have gained unprecedented prominence, achieving widespread adoption across diverse domains and integrating deeply into society. The capability to fine-tune general-purpose LLMs, such as Generative Pre-trained Transformers (GPT), for specific tasks has facilitated the emergence of numerous Custom GPTs. These tailored models are increasingly made available through dedicated marketplaces, such as OpenAI's GPT Store. However, their black-box nature introduces significant safety and compliance risks. In this work, we present a scalable framework for the automated evaluation of Custom GPTs against OpenAI's usage policies, which define the permissible behaviors of these systems. Our framework integrates three core components: (1) automated discovery and data collection of models from the GPT store, (2) a red-teaming prompt generator tailored to specific policy categories and the characteristics of each target GPT, and (3) an LLM-as-a-judge technique to analyze each prompt-response pair for potential policy violations. We validate our framework with a manually annotated ground truth, and evaluate it through a large-scale study with 782 Custom GPTs across three categories: Romantic, Cybersecurity, and Academic GPTs. Our manual annotation process achieved an F1 score of 0.975 in identifying policy violations, confirming the reliability of the framework's assessments. The results reveal that 58.7% of the analyzed models exhibit indications of non-compliance, exposing weaknesses in the GPT store's review and approval processes. Furthermore, our findings indicate that a model's popularity does not correlate with compliance, and non-compliance issues largely stem from behaviors inherited from base models rather than user-driven customizations. We believe this approach is extendable to other chatbot platforms and policy domains, improving LLM-based systems safety.
- Europe > Spain > Galicia > Madrid (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- Europe > United Kingdom > England > Greater London > London (0.04)
- Asia > Myanmar > Tanintharyi Region > Dawei (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study > Negative Result (0.46)
- Information Technology > Security & Privacy (1.00)
- Education (1.00)
- Government > Military > Cyberwarfare (0.35)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.50)
Are premium AI chatbots worth paying for? Only if you do these things
The future is coming, and that means making sure you're well-versed in AI usage for when it matters. And while there are all kinds of AI tools now, the most prominent one remains AI chatbots like ChatGPT. When I take a step back and look at my own AI chatbot use, that's what I'm predominantly doing with it: educating myself. Sure, I'm learning new skills and finding answers to questions when my kids stump me, but ultimately what I'm doing when I play around with the likes of ChatGPT Plus and Copilot Pro is learning how to use AI, period. And now that these AI chatbots have come out with premium subscription plans with even more advanced features, I wouldn't blame you if you're confused about why they have paid plans, what the benefits are, and whether they're worth paying for.
TalkMosaic: Interactive PhotoMosaic with Multi-modal LLM Q&A Interactions
We use images of cars of a wide range of varieties to compose an image of an animal such as a bird or a lion for the theme of environmental protection to maximize the information about cars in a single composed image and to raise the awareness about environmental challenges. We present a novel way of image interaction with an artistically-composed photomosaic image, in which a simple operation of "click and display" is used to demonstrate the interactive switch between a tile image in a photomosaic image and the corresponding original car image, which will be automatically saved on the Desktop. We build a multimodal custom GPT named TalkMosaic by incorporating car images information and the related knowledge to ChatGPT. By uploading the original car image to TalkMosaic, we can ask questions about the given car image and get the corresponding answers efficiently and effectively such as where to buy the tire in the car image that satisfies high environmental standards. We give an in-depth analysis on how to speed up the inference of multimodal LLM using sparse attention and quantization techniques with presented probabilistic FlashAttention (PrFlashAttention) and Staircase Adaptive Quantization (SAQ) methods. The implemented prototype demonstrates the feasibility and effectiveness of the presented approach.
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- Africa (0.04)
Leveraging LLMs for MT in Crisis Scenarios: a blueprint for low-resource languages
In an evolving landscape of crisis communication, the need for robust and adaptable Machine Translation (MT) systems is more pressing than ever, particularly for low-resource languages. This study presents a comprehensive exploration of leveraging Large Language Models (LLMs) and Multilingual LLMs (MLLMs) to enhance MT capabilities in such scenarios. By focusing on the unique challenges posed by crisis situations where speed, accuracy, and the ability to handle a wide range of languages are paramount, this research outlines a novel approach that combines the cutting-edge capabilities of LLMs with fine-tuning techniques and community-driven corpus development strategies. At the core of this study is the development and empirical evaluation of MT systems tailored for two low-resource language pairs, illustrating the process from initial model selection and fine-tuning through to deployment. Bespoke systems are developed and modelled on the recent Covid-19 pandemic. The research highlights the importance of community involvement in creating highly specialised, crisis-specific datasets and compares custom GPTs with NLLB-adapted MLLM models. It identifies fine-tuned MLLM models as offering superior performance compared with their LLM counterparts. A scalable and replicable model for rapid MT system development in crisis scenarios is outlined. Our approach enhances the field of humanitarian technology by offering a blueprint for developing multilingual communication systems during emergencies.
- Europe > United Kingdom > Northern Ireland (0.04)
- North America > United States > Pennsylvania > Philadelphia County > Philadelphia (0.04)
- North America > United States > Oregon > Washington County > Forest Grove (0.04)
- (8 more...)
- Research Report (1.00)
- Overview > Innovation (0.34)
- Health & Medicine > Therapeutic Area > Infections and Infectious Diseases (0.34)
- Health & Medicine > Therapeutic Area > Immunology (0.34)
- Health & Medicine > Epidemiology (0.34)
- Information Technology > Artificial Intelligence > Natural Language > Machine Translation (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
The application of GPT-4 in grading design university students' assignment and providing feedback: An exploratory study
Huang, Qian, Willems, Thijs, Poon, King Wang
This study aims to investigate whether GPT-4 can effectively grade assignments for design university students and provide useful feedback. In design education, assignments do not have a single correct answer and often involve solving an open-ended design problem. This subjective nature of design projects often leads to grading problems,as grades can vary between different raters,for instance instructor from engineering background or architecture background. This study employs an iterative research approach in developing a Custom GPT with the aim of achieving more reliable results and testing whether it can provide design students with constructive feedback. The findings include: First,through several rounds of iterations the inter-reliability between GPT and human raters reached a level that is generally accepted by educators. This indicates that by providing accurate prompts to GPT,and continuously iterating to build a Custom GPT, it can be used to effectively grade students' design assignments, serving as a reliable complement to human raters. Second, the intra-reliability of GPT's scoring at different times is between 0.65 and 0.78. This indicates that, with adequate instructions, a Custom GPT gives consistent results which is a precondition for grading students. As consistency and comparability are the two main rules to ensure the reliability of educational assessment, this study has looked at whether a Custom GPT can be developed that adheres to these two rules. We finish the paper by testing whether Custom GPT can provide students with useful feedback and reflecting on how educators can develop and iterate a Custom GPT to serve as a complementary rater.
- Asia > Singapore (0.05)
- Europe > Montenegro (0.04)
- North America > United States > Arizona (0.04)
- (2 more...)
- Instructional Material (1.00)
- Research Report > New Finding (0.68)
- Education > Educational Setting > Higher Education (1.00)
- Education > Assessment & Standards (1.00)
- Education > Educational Technology > Educational Software > Computer Based Training (0.46)
Efficacy of Large Language Models in Systematic Reviews
Shah, Aaditya, Mehendale, Shridhar, Kanthi, Siddha
This study investigates the effectiveness of Large Language Models (LLMs) in interpreting existing literature through a systematic review of the relationship between Environmental, Social, and Governance (ESG) factors and financial performance. The primary objective is to assess how LLMs can replicate a systematic review on a corpus of ESG-focused papers. We compiled and hand-coded a database of 88 relevant papers published from March 2020 to May 2024. Additionally, we used a set of 238 papers from a previous systematic review of ESG literature from January 2015 to February 2020. We evaluated two current state-of-the-art LLMs, Meta AI's Llama 3 8B and OpenAI's GPT-4o, on the accuracy of their interpretations relative to human-made classifications on both sets of papers. We then compared these results to a "Custom GPT" and a fine-tuned GPT-4o Mini model using the corpus of 238 papers as training data. The fine-tuned GPT-4o Mini model outperformed the base LLMs by 28.3% on average in overall accuracy on prompt 1. At the same time, the "Custom GPT" showed a 3.0% and 15.7% improvement on average in overall accuracy on prompts 2 and 3, respectively. Our findings reveal promising results for investors and agencies to leverage LLMs to summarize complex evidence related to ESG investing, thereby enabling quicker decision-making and a more efficient market.
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > New York > Kings County > New York City (0.04)
- North America > United States > Illinois (0.04)
- Europe > France > Normandy > Seine-Maritime > Rouen (0.04)
OpenAI's Custom Chatbots Are Leaking Their Secrets
You don't need to know how to code to create your own AI chatbot. Since the start of November--shortly before the chaos at the company unfolded--OpenAI has let anyone build and publish their own custom versions of ChatGPT, known as "GPTs". Thousands have been created: A "nomad" GPT gives advice about working and living remotely, another claims to search 200 million academic papers to answer your questions, and yet another will turn you into a Pixar character. However, these custom GPTs can also be forced into leaking their secrets. Security researchers and technologists probing the custom chatbots have made them spill the initial instructions they were given when they were created, and have also discovered and downloaded the files used to customize the chatbots.