Computer Speech and Language
Subject Area and Category
- Human-Computer Interaction
- Theoretical Computer Science
Academic Press
Publication type
08852308, 10958363
1986-1987, 1989-2024
Information
How to publish in this journal
The set of journals have been ranked according to their SJR and divided into four equal groups, four quartiles. Q1 (green) comprises the quarter of the journals with the highest values, Q2 (yellow) the second highest values, Q3 (orange) the third highest values and Q4 (red) the lowest values.
The SJR is a size-independent prestige indicator that ranks journals by their 'average prestige per article'. It is based on the idea that 'all citations are not created equal'. SJR is a measure of scientific influence of journals that accounts for both the number of citations received by a journal and the importance or prestige of the journals where such citations come from It measures the scientific influence of the average article in a journal, it expresses how central to the global scientific discussion an average article of the journal is.
Evolution of the number of published documents. All types of documents are considered, including citable and non citable documents.
This indicator counts the number of citations received by documents from a journal and divides them by the total number of documents published in that journal. The chart shows the evolution of the average number of times documents published in a journal in the past two, three and four years have been cited in the current year. The two years line is equivalent to journal impact factor ™ (Thomson Reuters) metric.
Evolution of the total number of citations and journal's self-citations received by a journal's published documents during the three previous years. Journal Self-citation is defined as the number of citation from a journal citing article to articles published by the same journal.
Evolution of the number of total citation per document and external citation per document (i.e. journal self-citations removed) received by a journal's published documents during the three previous years. External citations are calculated by subtracting the number of self-citations from the total number of citations received by the journal’s documents.
International Collaboration accounts for the articles that have been produced by researchers from several countries. The chart shows the ratio of a journal's documents signed by researchers from more than one country; that is including more than one country address.
Not every article in a journal is considered primary research and therefore "citable", this chart shows the ratio of a journal's articles including substantial research (research articles, conference papers and reviews) in three year windows vs. those documents other than research articles, reviews and conference papers.
Ratio of a journal's items, grouped in three years windows, that have been cited at least once vs. those not cited during the following year.
Evolution of the percentage of female authors.
Evolution of the number of documents cited by public policy documents according to Overton database.
Evoution of the number of documents related to Sustainable Development Goals defined by United Nations. Available from 2018 onwards.
Leave a comment
Name * Required
Email (will not be published) * Required
* Required Cancel
The users of Scimago Journal & Country Rank have the possibility to dialogue through comments linked to a specific journal. The purpose is to have a forum in which general doubts about the processes of publication in the journal, experiences and other issues derived from the publication of papers are resolved. For topics on particular articles, maintain the dialogue through the usual channels with your editor.
Follow us on @ScimagoJR Scimago Lab , Copyright 2007-2024. Data Source: Scopus®
Cookie settings
Cookie Policy
Legal Notice
Privacy Policy
- solidarity - (ua) - (ru)
- news - (ua) - (ru)
- donate - donate - donate
for scientists:
- ERA4Ukraine
- Assistance in Germany
- Ukrainian Global University
- #ScienceForUkraine
default search action
- combined dblp search
- author search
- venue search
- publication search
Computer Speech & Language
- > Home > Journals
Venue statistics
records by year
frequent authors
Venue Information
- issn: 0885-2308
Computer Speech & Language @ ScienceDirect
- 2024: Volumes 83 , 84 , 85 ,
- 2023: Volumes 77 , 78 , 79 , 80 , 81 , 82
- 2022: Volumes 71 , 72 , 73 , 74 , 75 , 76
- 2021: Volumes 66 , 67 , 68 , 69 , 70
- 2020: Volumes 59 , 60 , 61 , 62 , 63 , 64 , 65
- 2019: Volumes 53 , 54 , 55 , 56 , 57 , 58
- 2018: Volumes 47 , 48 , 49 , 50 , 51 , 52
- 2017: Volumes 41 , 42 , 43 , 44 , 45 , 46
- 2016: Volumes 35 , 36 , 37 , 38 , 39 , 40
- 2015: Volumes 30 , 31 , 32 , 33 , 34
- Volume 29: 2015
- Volume 28: 2014
- Volume 27: 2013
- Volume 26: 2012
- Volume 25: 2011
- Volume 24: 2010
- Volume 23: 2009
- Volume 22: 2008
- Volume 21: 2007
- Volume 20: 2006
- Volume 19: 2005
- Volume 18: 2004
- Volume 17: 2003
- Volume 16: 2002
- Volume 15: 2001
- Volume 14: 2000
- Volume 13: 1999
- Volume 12: 1998
- Volume 11: 1997
- Volume 10: 1996
- Volume 9: 1995
- Volume 8: 1994
- Volume 7: 1993
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default . You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
load links from unpaywall.org
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy .
Archived links via Wayback Machine
load content from archive.org
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy .
Reference lists
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org , opencitations.net , and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy , as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
load data from openalex.org
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex .
last updated on 2024-05-10 18:46 CEST by the dblp team
see also: Terms of Use | Privacy Policy | Imprint
dblp was originally created in 1993 at:
since 2018, dblp has been operated and maintained by:
the dblp computer science bibliography is funded and supported by:
Search form
Computer Speech And Language
You may order single or multiple copies of back and recent journal issues. If you are an Author wishing to obtain a printed copy of the journal issue featuring your article, or you require a printed copy for research, review or add to your library, the process is easy –
- Select your journal volume and issue.
- Select the required quantity in the Review cart page
- Provide the shipping details and process the payment.
- Average production time is approx. 2 weeks.
- Your shipping options and general shipping times are: DHL for international - 2-5 postal days and UPS for domestic – 1-6 business days depending on delivery address. . We can track your shipment status at any time.
International Journal of Speech Technology
The International Journal of Speech Technology is dedicated to promoting research in all aspects of speech input and output, including applications and base technology.
- Presents papers on the base technology and theory as well as the full spectrum of applications.
- Represents the only journal encompassing all aspects of the three major technologies: text-to-speech synthesis, automatic speech recognition, and stored (digitized) speech.
- A significant 92% of authors reported they would likely publish in the journal again.
- Explores technological issues of speech input or output, novel algorithms, and Natural Language Processing (NLP).
- Provides an international and local language implementation of speech synthesis and recognition, while giving consideration to under-resourced languages.
- Amy Neustein
Latest issue
Volume 27, Issue 1
Latest articles
Unsupervised phoneme segmentation of continuous arabic speech.
- Hind Ait Mait
- Noureddine Aboutabit
Improving low-complexity and real-time DeepFilterNet2 for personalized speech enhancement
- Shilin Wang
- Haixin Guan
- Yanhua Long
A computationally efficient speech emotion recognition system employing machine learning classifiers and ensemble learning
- N. Aishwarya
- Kanwaljeet Kaur
- Karthik Seemakurthy
Feature extraction using GTCC spectrogram and ResNet50 based classification for audio spoof detection
- Nidhi Chakravarty
Speech recognition based on the transformer's multi-head attention in Arabic
- Omayma Mahmoudi
- Mouncef Filali-Bouami
- Mohamed Benchat
Journal updates
Guest-edited collections policies.
Review the journal's special issue submission guidelines and restrictions here.
Contact the Journal
Questions about your submission to the journal? Find the right person to answer them by clicking here.
Journal information
- ACM Digital Library
- EI Compendex
- Google Scholar
- Japanese Science and Technology Agency (JST)
- Norwegian Register for Scientific Journals and Series
- OCLC WorldCat Discovery Service
- TD Net Discovery Service
- UGC-CARE List (India)
Rights and permissions
Editorial policies
© Springer Science+Business Media, LLC, part of Springer Nature
- Find a journal
- Publish with us
- Track your research
COMPUTER SPEECH AND LANGUAGE
- Journal Search
- Journal Details
Note: The following journal information is for reference only. Please check the journal website for updated information prior to submission.
COMPUT SPEECH LANG
COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
- Popular journals in the same field
- Recent articles
Verified Reviews
Find the ideal target journal for your manuscript.
Explore over 38,000 international journals covering a vast array of academic fields.
Create your own webinar
Interested in hosting your own webinar? Check the schedule and propose your idea to the Peeref Content Team.
Computer Speech and Language
( API-Link )
Impact Factor : 4.300 (based on Web of Science 2022)
- # 44 / 131 (Q2) in Computer Science, Artificial Intelligence
Partner: • University Press Alert
- Language Editing     For Manuscripts    For Response Letter new    For LaTeX    For Annual Review and Tenure    For Books new
- Scientific Editing     For Manuscripts    For Response Letter new
- Grant Editing 
- Translation 
- Publication Support  Journal Recommendation  Manuscript Formatting  Figure Formatting  Data Analysis new  Plagiarism Check  Conference Poster  Plain Language Summary
- Scientific Illustration  Journal Cover Design  Graphical Abstract  Infographic  Custom Illustration
- Scientific Videos  Video Abstract  Explainer Video  Scientific Animation
- Ethics and Confidentiality
- Editorial Certificate
- Testimonials
- Design Gallery
- Institutional Provider
- Publisher Portal
- Brand Localization
- Journal Selector Tool
- Learning Nexus
Scientific Journal Selector
COMPUTER SPEECH AND LANGUAGE
APA has partnered with LetPub to provide a full suite of author services.
Free Webinar Series Conversations with a Hindawi Editor
Professional Journal Cover Design
Professionally designed and impactful journal cover art. Delivered fast and consistent with journal guidelines.
Intentional Space Tag
Contact us
Your name *
Your email *
Your message *
Please fill in all fields and provide a valid email.
© 2010-2024 ACCDON LLC 400 5 th Ave, Suite 530, Waltham, MA 02451, USA Privacy • Terms of Service
© 2010-2024 United States: ACCDON LLC Tel: 1-781-202-9968 Email: [email protected]
Address: 400 5 th Ave, Suite 530, Waltham, Massachusetts 02451, United States
NEWS: Chatbots.org survey on 3000 US and UK consumers shows it is time for chatbot integration in customer service! read more..
Computer Speech and Language
Summary: Computer Speech and Language covers all topics concerning speech and language.
Computer Speech & Language publishes reports of original research related to the recognition, understanding, production, coding and mining of speech and language.
The speech and language sciences have a long history, but it is only relatively recently that large-scale implementation of and experimentation with complex models of speech and language processing has become feasible. Such research is often carried out somewhat separately by practitioners of artificial intelligence, computer science, electronic engineering, information retrieval, linguistics, phonetics, or psychology.
The journal provides a focus for this work, and encourages an interdisciplinary approach to speech and language research and technology. Thus contributions from all of the related fields are welcomed in the form of reports of theoretical or experimental studies, tutorials, and brief correspondence pertaining to models and their implementation, or reports of fundamental research leading to the improvement of such models.
New Comment
Search Journals
Browse topics.
- Virtual reality
- Virtual worlds
- Translation
- Speech synthesis (TTS)
- Text recognition
- User Client Technology
- Face recognition
- Fingerprint recognition
- Gait analysis
- Iris recognition
- Typing rhythm recognition
Most Relevant Journals
- Interdisciplinary Advances in Adaptive and Intelligent Assistant Systems
- International Journal of Virtual Technology and Multimedia
- Agent Technologies and Web Engineering: Applications and Systems
- Computer Animation and Virtual Worlds
- Web Intelligence and Agent Systems: An International Journal
- Journal of Artificial Intelligence
- Journal of Intelligent and Robotic Systems
- International Journal of Intelligent Systems
- Artificial Intelligence
- Applied Artificial Intelligence
Just added in our library
Comments in research sections.
it is a nice academic journa, i like to publish papers in it.
Speaking of cancer, the air Force could spray united states from above with a wonderfully "safe" …
I've just watched this video and I gotta say: this is really a big breakthrough! I …
This article features a storytelling application for mobile devices that uses an anthropomorphic guide, Carletto. Carletto …
Thanks for this extensive review Mark-Shane!!!! This is what we REALLY need in this industry!! As …
In our business section
- BRAIN: A Platform That Enables Businesses To Create Chatbots And More
- An International Chatbots.org Survey: Consumers say No to Chatbot Silos
- Chatbot which recognizes objects
- Come and join me at UTTR in London Oct 3th!
- Cool chatbot animation
Hot on AI Zone
AI forum hosted by Chatbots.org
- More access by clicking "Chat Now"
- Tom Walker? (This one is readable)
- What is going on with your site?
- CS and GPT-x
- Compiling basic question responses for about 2000 common English nouns
Browse All Chatbot Categories
- Sales Enablement Chat
- Real Estate Chatbots
- Ecommerce Chatbots
- Messenger Chat
- Customer Support Chat
- Live Chat Integration
- Enterprise Chatbots
- SMS Marketing Services
- Chatbots for Marketing
- Chatbot Builders
Chatbot Reviews
Latest Chatbot Reviews
- MobileMonkey Review
- ManyChat Review
- Chatfuel Review
- ItsAlive Review
- ChatterOn Review
- Botsociety Review
- Intercom Review
- Drift Review
- Zendesk Chat Review
- EZ Texting Review
- Dialogflow Review
- Olark Review
- Freshchat Review
- Pandorabots Review
- Flow XO Review
- Bold360 Review
- HubSpot Conversations Review
- Octane AI Review
- Botsify Review
- SimpleTexting Review
- Acquire io Review
Use our Chat Match Tool to get started with Chatbots for Business
Compare features, pricing, and reviews from award-winning providers based on best fit for your business., how many team members (marketing, sales, it and customer support) will be involved in your chatbot system.
This info helps us provide matches to platforms designed for your company's size.
Please choose one option
1-5 6-25 26-50 50+
We're putting your report together.
Where should we send it?
Please enter a valid Email address
What chat automation functions are most important to you? Check all that apply.
This will help us match you to providers that cater to your specific needs.
Marketing agency friendly: multi-client, multi-user Lead gen & customer acquisition Customer retention, engagement & remarketing Native to website Native to mobile app Native to Facebook Customer support automation + live chat hand-off Business system integration (order fulfillment, CRM, etc.) Other
Who should we send the information to?
What is your business website?
Please enter a valid url
What is the best number to reach you?
Your privacy is our priority.
Please enter a valid Phone Number
Your request has been received. A dedicated specialist will contact you shortly to provide you with free pricing information. For an immediate call back, please email [email protected] .
Computer Speech and Language Impact Factor & Key Scientometrics
Computer speech and language overview, impact factor.
I. Basic Journal Info
Journal ISSN: 08852308, 10958363
Publisher: elsevier inc., history: 1986-1987, 1989-ongoing, journal hompage: link, how to get published:, research categories, scope/description:.
--------------------------------
Best Academic Tools
- Academic Writing Tools
- Proofreading Tools
- Academic Search Engines
- Project Management Tools
- Survey Tools for Research
- Transcription Tools
- Reference Management Software
- AI-Based Summary Generators
- Academic Social Network Sites
- Plagiarism Checkers
- Science Communication Tools
- Jasper AI Review
II. Science Citation Report (SCR)
Computer speech and language scr impact factor, computer speech and language scr journal ranking, computer speech and language scimago sjr rank.
SCImago Journal Rank (SJR indicator) is a measure of scientific influence of scholarly journals that accounts for both the number of citations received by a journal and the importance or prestige of the journals where such citations come from.
Computer Speech and Language Scopus 2-Year Impact Factor Trend
Computer speech and language scopus 3-year impact factor trend, computer speech and language scopus 4-year impact factor trend, computer speech and language impact factor history.
- 2022 Impact Factor 5.385 4.703 4.315
- 2021 Impact Factor 4.479 3.95 4.044
- 2020 Impact Factor 2.73 3.262 2.922
- 2019 Impact Factor 3.901 3.382 3.415
- 2018 Impact Factor 2.862 3.019 3.121
- 2017 Impact Factor 2.975 3.217 3.041
- 2016 Impact Factor 3.13 3.068 2.956
- 2015 Impact Factor 3.253 3.154 3.192
- 2014 Impact Factor 2.615 NA NA
- 2013 Impact Factor 3.134 NA NA
- 2012 Impact Factor 3.86 NA NA
- 2011 Impact Factor 3.729 NA NA
- 2010 Impact Factor 3.02 NA NA
- 2009 Impact Factor 2.672 NA NA
- 2008 Impact Factor 3.569 NA NA
- 2007 Impact Factor 3 NA NA
- 2006 Impact Factor 1.766 NA NA
- 2005 Impact Factor 1.22 NA NA
- 2004 Impact Factor 2.488 NA NA
- 2003 Impact Factor 2.146 NA NA
- 2002 Impact Factor 1.167 NA NA
- 2001 Impact Factor 1.541 NA NA
- 2000 Impact Factor 1.659 NA NA
See what other people are reading
HIGHEST PAID JOBS
- Highest Paying Nursing Jobs
- Highest Paying Non-Physician Jobs
- Highest Paying Immunology Jobs
- Highest Paying Microbiology Jobs
LATEX TUTORIALS
- LaTeX Installation Guide – Easy to Follow Steps to Install LaTeX
- 6 Easy Steps to Create Your First LaTeX Document
- How to Use LaTeX Paragraphs and Sections
- How to Use LaTeX Packages with Examples
MUST-READ BOOKS
- Multidisciplinary
- Health Science
Impact factor (IF) is a scientometric factor based on the yearly average number of citations on articles published by a particular journal in the last two years. A journal impact factor is frequently used as a proxy for the relative importance of a journal within its field. Find out more: What is a good impact factor?
III. Other Science Influence Indicators
Any impact factor or scientometric indicator alone will not give you the full picture of a science journal. There are also other factors such as H-Index, Self-Citation Ratio, SJR, SNIP, etc. Researchers may also consider the practical aspect of a journal such as publication fees, acceptance rate, review speed. ( Learn More )
Computer Speech and Language H-Index
The h-index is an author-level metric that attempts to measure both the productivity and citation impact of the publications of a scientist or scholar. The index is based on the set of the scientist's most cited papers and the number of citations that they have received in other publications
Computer Speech and Language H-Index History
scijournal.org is a platform dedicated to making the search and use of impact factors of science journals easier.
Suggestions or feedback?
MIT News | Massachusetts Institute of Technology
- Machine learning
- Social justice
- Black holes
- Classes and programs
Departments
- Aeronautics and Astronautics
- Brain and Cognitive Sciences
- Architecture
- Political Science
- Mechanical Engineering
Centers, Labs, & Programs
- Abdul Latif Jameel Poverty Action Lab (J-PAL)
- Picower Institute for Learning and Memory
- Lincoln Laboratory
- School of Architecture + Planning
- School of Engineering
- School of Humanities, Arts, and Social Sciences
- Sloan School of Management
- School of Science
- MIT Schwarzman College of Computing
Natural language boosts LLM performance in coding, planning, and robotics
Press contact :.
Previous image Next image
Large language models (LLMs) are becoming increasingly useful for programming and robotics tasks, but for more complicated reasoning problems, the gap between these systems and humans looms large. Without the ability to learn new concepts like humans do, these systems fail to form good abstractions — essentially, high-level representations of complex concepts that skip less-important details — and thus sputter when asked to do more sophisticated tasks. Luckily, MIT Computer Science and Artificial Intelligence Laboratory (CSAIL) researchers have found a treasure trove of abstractions within natural language. In three papers to be presented at the International Conference on Learning Representations this month, the group shows how our everyday words are a rich source of context for language models, helping them build better overarching representations for code synthesis, AI planning, and robotic navigation and manipulation. The three separate frameworks build libraries of abstractions for their given task: LILO (library induction from language observations) can synthesize, compress, and document code; Ada (action domain acquisition) explores sequential decision-making for artificial intelligence agents; and LGA (language-guided abstraction) helps robots better understand their environments to develop more feasible plans. Each system is a neurosymbolic method, a type of AI that blends human-like neural networks and program-like logical components. LILO: A neurosymbolic framework that codes Large language models can be used to quickly write solutions to small-scale coding tasks, but cannot yet architect entire software libraries like the ones written by human software engineers. To take their software development capabilities further, AI models need to refactor (cut down and combine) code into libraries of succinct, readable, and reusable programs. Refactoring tools like the previously developed MIT-led Stitch algorithm can automatically identify abstractions, so, in a nod to the Disney movie “Lilo & Stitch,” CSAIL researchers combined these algorithmic refactoring approaches with LLMs. Their neurosymbolic method LILO uses a standard LLM to write code, then pairs it with Stitch to find abstractions that are comprehensively documented in a library. LILO’s unique emphasis on natural language allows the system to do tasks that require human-like commonsense knowledge, such as identifying and removing all vowels from a string of code and drawing a snowflake. In both cases, the CSAIL system outperformed standalone LLMs, as well as a previous library learning algorithm from MIT called DreamCoder, indicating its ability to build a deeper understanding of the words within prompts. These encouraging results point to how LILO could assist with things like writing programs to manipulate documents like Excel spreadsheets, helping AI answer questions about visuals, and drawing 2D graphics.
“Language models prefer to work with functions that are named in natural language,” says Gabe Grand SM '23, an MIT PhD student in electrical engineering and computer science, CSAIL affiliate, and lead author on the research. “Our work creates more straightforward abstractions for language models and assigns natural language names and documentation to each one, leading to more interpretable code for programmers and improved system performance.”
When prompted on a programming task, LILO first uses an LLM to quickly propose solutions based on data it was trained on, and then the system slowly searches more exhaustively for outside solutions. Next, Stitch efficiently identifies common structures within the code and pulls out useful abstractions. These are then automatically named and documented by LILO, resulting in simplified programs that can be used by the system to solve more complex tasks.
The MIT framework writes programs in domain-specific programming languages, like Logo, a language developed at MIT in the 1970s to teach children about programming. Scaling up automated refactoring algorithms to handle more general programming languages like Python will be a focus for future research. Still, their work represents a step forward for how language models can facilitate increasingly elaborate coding activities. Ada: Natural language guides AI task planning Just like in programming, AI models that automate multi-step tasks in households and command-based video games lack abstractions. Imagine you’re cooking breakfast and ask your roommate to bring a hot egg to the table — they’ll intuitively abstract their background knowledge about cooking in your kitchen into a sequence of actions. In contrast, an LLM trained on similar information will still struggle to reason about what they need to build a flexible plan. Named after the famed mathematician Ada Lovelace, who many consider the world’s first programmer, the CSAIL-led “Ada” framework makes headway on this issue by developing libraries of useful plans for virtual kitchen chores and gaming. The method trains on potential tasks and their natural language descriptions, then a language model proposes action abstractions from this dataset. A human operator scores and filters the best plans into a library, so that the best possible actions can be implemented into hierarchical plans for different tasks. “Traditionally, large language models have struggled with more complex tasks because of problems like reasoning about abstractions,” says Ada lead researcher Lio Wong, an MIT graduate student in brain and cognitive sciences, CSAIL affiliate, and LILO coauthor. “But we can combine the tools that software engineers and roboticists use with LLMs to solve hard problems, such as decision-making in virtual environments.”
When the researchers incorporated the widely-used large language model GPT-4 into Ada, the system completed more tasks in a kitchen simulator and Mini Minecraft than the AI decision-making baseline “Code as Policies.” Ada used the background information hidden within natural language to understand how to place chilled wine in a cabinet and craft a bed. The results indicated a staggering 59 and 89 percent task accuracy improvement, respectively. With this success, the researchers hope to generalize their work to real-world homes, with the hopes that Ada could assist with other household tasks and aid multiple robots in a kitchen. For now, its key limitation is that it uses a generic LLM, so the CSAIL team wants to apply a more powerful, fine-tuned language model that could assist with more extensive planning. Wong and her colleagues are also considering combining Ada with a robotic manipulation framework fresh out of CSAIL: LGA (language-guided abstraction). Language-guided abstraction: Representations for robotic tasks Andi Peng SM ’23, an MIT graduate student in electrical engineering and computer science and CSAIL affiliate, and her coauthors designed a method to help machines interpret their surroundings more like humans, cutting out unnecessary details in a complex environment like a factory or kitchen. Just like LILO and Ada, LGA has a novel focus on how natural language leads us to those better abstractions. In these more unstructured environments, a robot will need some common sense about what it’s tasked with, even with basic training beforehand. Ask a robot to hand you a bowl, for instance, and the machine will need a general understanding of which features are important within its surroundings. From there, it can reason about how to give you the item you want.
In LGA’s case, humans first provide a pre-trained language model with a general task description using natural language, like “bring me my hat.” Then, the model translates this information into abstractions about the essential elements needed to perform this task. Finally, an imitation policy trained on a few demonstrations can implement these abstractions to guide a robot to grab the desired item. Previous work required a person to take extensive notes on different manipulation tasks to pre-train a robot, which can be expensive. Remarkably, LGA guides language models to produce abstractions similar to those of a human annotator, but in less time. To illustrate this, LGA developed robotic policies to help Boston Dynamics’ Spot quadruped pick up fruits and throw drinks in a recycling bin. These experiments show how the MIT-developed method can scan the world and develop effective plans in unstructured environments, potentially guiding autonomous vehicles on the road and robots working in factories and kitchens.
“In robotics, a truth we often disregard is how much we need to refine our data to make a robot useful in the real world,” says Peng. “Beyond simply memorizing what’s in an image for training robots to perform tasks, we wanted to leverage computer vision and captioning models in conjunction with language. By producing text captions from what a robot sees, we show that language models can essentially build important world knowledge for a robot.” The challenge for LGA is that some behaviors can’t be explained in language, making certain tasks underspecified. To expand how they represent features in an environment, Peng and her colleagues are considering incorporating multimodal visualization interfaces into their work. In the meantime, LGA provides a way for robots to gain a better feel for their surroundings when giving humans a helping hand.
An “exciting frontier” in AI
“Library learning represents one of the most exciting frontiers in artificial intelligence, offering a path towards discovering and reasoning over compositional abstractions,” says assistant professor at the University of Wisconsin-Madison Robert Hawkins, who was not involved with the papers. Hawkins notes that previous techniques exploring this subject have been “too computationally expensive to use at scale” and have an issue with the lambdas, or keywords used to describe new functions in many languages, that they generate. “They tend to produce opaque 'lambda salads,' big piles of hard-to-interpret functions. These recent papers demonstrate a compelling way forward by placing large language models in an interactive loop with symbolic search, compression, and planning algorithms. This work enables the rapid acquisition of more interpretable and adaptive libraries for the task at hand.” By building libraries of high-quality code abstractions using natural language, the three neurosymbolic methods make it easier for language models to tackle more elaborate problems and environments in the future. This deeper understanding of the precise keywords within a prompt presents a path forward in developing more human-like AI models. MIT CSAIL members are senior authors for each paper: Joshua Tenenbaum, a professor of brain and cognitive sciences, for both LILO and Ada; Julie Shah, head of the Department of Aeronautics and Astronautics, for LGA; and Jacob Andreas, associate professor of electrical engineering and computer science, for all three. The additional MIT authors are all PhD students: Maddy Bowers and Theo X. Olausson for LILO, Jiayuan Mao and Pratyusha Sharma for Ada, and Belinda Z. Li for LGA. Muxin Liu of Harvey Mudd College was a coauthor on LILO; Zachary Siegel of Princeton University, Jaihai Feng of the University of California at Berkeley, and Noa Korneev of Microsoft were coauthors on Ada; and Ilia Sucholutsky, Theodore R. Sumers, and Thomas L. Griffiths of Princeton were coauthors on LGA. LILO and Ada were supported, in part, by MIT Quest for Intelligence, the MIT-IBM Watson AI Lab, Intel, U.S. Air Force Office of Scientific Research, the U.S. Defense Advanced Research Projects Agency, and the U.S. Office of Naval Research, with the latter project also receiving funding from the Center for Brains, Minds and Machines. LGA received funding from the U.S. National Science Foundation, Open Philanthropy, the Natural Sciences and Engineering Research Council of Canada, and the U.S. Department of Defense.
Share this news article on:
Related links.
- Jacob Andreas
- Computer Science and Artificial Intelligence Laboratory (CSAIL)
- MIT Language and Intelligence Group
- Center for Brains, Minds, and Machines
- Department of Electrical Engineering and Computer Science
- Department of Brain and Cognitive Sciences
- Department of Aeronautics and Astronautics
Related Topics
- Computer science and technology
- Artificial intelligence
- Natural language processing
- Electrical Engineering & Computer Science (eecs)
- Programming
- Human-computer interaction
- Computer vision
- Brain and cognitive sciences
- Programming languages
- Center for Brains Minds and Machines
- Quest for Intelligence
- MIT-IBM Watson AI Lab
- National Science Foundation (NSF)
- Department of Defense (DoD)
- Defense Advanced Research Projects Agency (DARPA)
Related Articles
Large language models use a surprisingly simple mechanism to retrieve some stored knowledge
Reasoning and reliability in AI
AI agents help explain other AI systems
3 Questions: Jacob Andreas on large language models
Previous item Next item
More MIT News
The power of App Inventor: Democratizing possibilities for mobile applications
Read full story →
Using MRI, engineers have found a way to detect light deep in the brain
A better way to control shape-shifting soft robots
From steel engineering to ovarian tumor research
Professor Emeritus David Lanning, nuclear engineer and key contributor to the MIT Reactor, dies at 96
Discovering community and cultural connections
- More news on MIT News homepage →
Massachusetts Institute of Technology 77 Massachusetts Avenue, Cambridge, MA, USA
- Map (opens in new window)
- Events (opens in new window)
- People (opens in new window)
- Careers (opens in new window)
- Accessibility
- Social Media Hub
- MIT on Facebook
- MIT on YouTube
- MIT on Instagram
IMAGES
VIDEO
COMMENTS
About the journal. Computer Speech & Language publishes reports of original research related to the recognition, understanding, production, coding and mining of speech and language. The speech and language sciences have a long history, but it is only relatively recently that large-scale implementation of and experimentation with complex models ...
Deep learning-based speech deception detection research relies on a large amount of labeled data. However, in the process of collecting speech deception detection data, the identification of truth and lies requires researchers to have a ... Highlights . A new semi supervised speech deception detection algorithm.
Computer Speech and Language. Search within CSPL. Search Search. Home; Browse by Title; Periodicals; Computer Speech and Language; Computer Speech and Language. Volume 85, Issue C. Apr 2024. Read More. Academic Press Ltd. 24-28 Oval Rd. London NW1 7DX; United Kingdom; Get Alerts for this Periodical Alerts. Share on.
Corrigendum to 'Unsupervised sign language validation process based on hand-motion parameter clustering' <Computer Speech & Language Volume 71, January 2022, 101256>. Mehrez Boulares, Ahmed Barnawi. Article 101319.
Scope. Computer Speech & Language publishes reports of original research related to the recognition, understanding, production, coding and mining of speech and language. The speech and language sciences have a long history, but it is only relatively recently that large-scale implementation of and experimentation with complex models of speech ...
Computer Speech & Language publishes reports of original research related to the recognition, understanding, production, coding and mining of speech and language. ... The journal provides a focus for this work, and encourages an interdisciplinary approach to speech and language research and technology. Thus contributions from all of the related ...
Language and Speech is a peer-reviewed journal which provides an international forum for communication among researchers in the disciplines that contribute to our understanding of human production, perception, processing, learning, use, and disorders of speech and language. The journal accepts reports of original research in all these areas. Interdisciplinary submissions are e
Bibliographic content of Computer Speech & Language. Stop the war! Остановите войну! solidarity - - news - - donate - donate - donate; for scientists: ERA4Ukraine; Assistance in Germany; ... the dblp computer science bibliography is funded and supported by: ...
Know all about Computer Speech and Language - Impact factor, Acceptance rate, Scite Analysis, H-index, SNIP Score, ISSN, Citescore, SCImago Journal Ranking (SJR), Aims & Scope, Publisher, and Other Important Metrics. Click to know more about Computer Speech and Language Review Speed, Scope, Publication Fees, Submission Guidelines.
Computer Speech And Language. You may order single or multiple copies of back and recent journal issues. If you are an Author wishing to obtain a printed copy of the journal issue featuring your article, or you require a printed copy for research, review or add to your library, the process is easy -. Select your journal volume and issue.
The ISSN of Computer Speech and Language journal is 08852308, 10958363. An International Standard Serial Number (ISSN) is a unique code of 8 digits. It is used for the recognition of journals, newspapers, periodicals, and magazines in all kind of forms, be it print-media or electronic.
The International Journal of Speech Technology is dedicated to promoting research in all aspects of speech input and output, including applications and base technology. Presents papers on the base technology and theory as well as the full spectrum of applications. Represents the only journal encompassing all aspects of the three major ...
In another review, the authors (2022b) investigated 57 publications in 10 journals until 2019 and identified mobile devices, multimedia, speech-to-text and text-to-speech, and digital game-based learning as the top five technologies used in language teaching.
An official publication of the International Speech Communication Association (ISCA) Computer Speech & Language publishes reports of original research related to the recognition, understanding, production, coding and mining of speech and language. The speech and language sciences have a long history, but it is only relatively recently that large-scale implementation of and experimentation with ...
Open data-based citation metrics about Computer Speech and Language, but also research trends, citation patterns, altmetric scores, similar journals and impact factors. ... About; Computer Speech and Language. Journal Metrics (Based on the publications from the last 4 years) (from 2019-03-01 to roughly 2023-03-01) Number of papers: 100: H4 ...
0885-2308 / 1095-8363. Aims and Scope. Computer Speech & Language publishes reports of original research related to the recognition, understanding, production, coding and mining of speech and language. The speech and language sciences have a long history, but it is only relatively recently that large-scale implementation of and experimentation ...
Open data-based citation metrics about Computer Speech and Language, but also research trends, citation patterns, altmetric scores, similar journals and impact factors. ... (Based on citations to the other journals in the most recent 30 papers in this journal, at least if metadata about citations were available; last updated 2023-04-12 ...
COMPUTER SPEECH AND LANGUAGE ... Subcategory: Theoretical Computer Science: 12 / 127: Category: Mathematics Subcategory: Software: 66 / 404: ... Journal Cover Design. Professionally designed and impactful journal cover art. Delivered fast and consistent with journal guidelines.
Computer Speech & Language publishes reports of original research related to the recognition, understanding, production, coding and mining of speech and language. ... The journal provides a focus for this work, and encourages an interdisciplinary approach to speech and language research and technology. Thus contributions from all of the related ...
Scope/Description: Computer Speech & Language publishes reports of original research related to the recognition, understanding, production, coding and mining of speech and language. The speech and language sciences have a long history, but it is only relatively recently that large-scale implementation of and experimentation with complex models ...
Computer Speech & Language focuses largely on the fields of Artificial intelligence, Speech recognition, Natural language processing, Artificial neural network and Language model. Artificial intelligence research featured in the journal incorporates concerns from various other topics such as Key (cryptography) and Pattern recognition.
Scaling up automated refactoring algorithms to handle more general programming languages like Python will be a focus for future research. Still, their work represents a step forward for how language models can facilitate increasingly elaborate coding activities. Ada: Natural language guides AI task planning