{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,9]],"date-time":"2026-03-09T04:23:01Z","timestamp":1773030181001,"version":"3.50.1"},"reference-count":25,"publisher":"Association for Computing Machinery (ACM)","issue":"6","funder":[{"name":"FSE invest in your future.","award":["MCIN\/AEI\/10.13039\/501100011033"],"award-info":[{"award-number":["MCIN\/AEI\/10.13039\/501100011033"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Intell. Syst. Technol."],"published-print":{"date-parts":[[2025,12,31]]},"abstract":"<jats:p>Nowadays, Generative Large Language Models (GLLMs) have made a significant impact in the field of Artificial Intelligence (AI). One of the domains extensively explored for these models is their ability as generators of functional source code for software projects. Nevertheless, their potential as assistants to write the code needed to generate and model Machine Learning (ML) or Deep Learning (DL) architectures has not been fully explored to date. For this reason, this work focuses on evaluating the extent to which different tools based on GLLMs, such as ChatGPT or Copilot, are able to correctly define the source code necessary to generate viable predictive models. The use case defined is the forecasting of a time series that reports the indoor temperature of a greenhouse. The results indicate that, while it is possible to achieve good accuracy metrics with simple predictive models generated by GLLMs, the composition of predictive models with complex architectures using GLLMs is still far from improving the accuracy of predictive models generated by human data scientists.<\/jats:p>","DOI":"10.1145\/3663485","type":"journal-article","created":{"date-parts":[[2024,5,7]],"date-time":"2024-05-07T15:27:24Z","timestamp":1715095644000},"page":"1-12","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":5,"title":["Developing Time Series Forecasting Models with Generative Large Language Models"],"prefix":"10.1145","volume":"16","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0008-4825","authenticated-orcid":false,"given":"Juan","family":"Morales-Garc\u00eda","sequence":"first","affiliation":[{"name":"Universidad Cat\u00f3lica de Murcia, Murcia, Spain"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9802-4240","authenticated-orcid":false,"given":"Antonio","family":"Llanes","sequence":"additional","affiliation":[{"name":"Universidad Cat\u00f3lica de Murcia, Murcia, Spain"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4892-5902","authenticated-orcid":false,"given":"Francisco","family":"Arcas-T\u00fanez","sequence":"additional","affiliation":[{"name":"Universidad Cat\u00f3lica de Murcia, Murcia, Spain"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1921-1137","authenticated-orcid":false,"given":"Fernando","family":"Terroso-S\u00e1enz","sequence":"additional","affiliation":[{"name":"Universidad Polit\u00e9cnica de Cartagena, Murcia, Spain"}]}],"member":"320","published-online":{"date-parts":[[2025,11,24]]},"reference":[{"key":"e_1_3_2_2_2","doi-asserted-by":"publisher","DOI":"10.1080\/07474938.2010.481556"},{"key":"e_1_3_2_3_2","doi-asserted-by":"publisher","DOI":"10.1080\/07474938.2010.481556"},{"key":"e_1_3_2_4_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICEngTechnol.2017.8308186"},{"key":"e_1_3_2_5_2","doi-asserted-by":"publisher","DOI":"10.1108\/EJIM-02-2023-0156"},{"key":"e_1_3_2_6_2","doi-asserted-by":"publisher","DOI":"10.1201\/9781420036206"},{"key":"e_1_3_2_7_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4419-9326-7_5"},{"key":"e_1_3_2_8_2","doi-asserted-by":"crossref","unstructured":"Ties de Kok. 2023. Generative LLMs and textual analysis in accounting:(Chat) GPT as research assistant? Available at SSRN.","DOI":"10.2139\/ssrn.4429658"},{"key":"e_1_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1145\/3597926.3598067"},{"key":"e_1_3_2_10_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.iotcps.2023.05.004"},{"key":"e_1_3_2_11_2","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"e_1_3_2_12_2","doi-asserted-by":"publisher","DOI":"10.1145\/3474349.3480209"},{"key":"e_1_3_2_13_2","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3501870"},{"key":"e_1_3_2_14_2","doi-asserted-by":"publisher","DOI":"10.1038\/nature14539"},{"key":"e_1_3_2_15_2","first-page":"2194","article-title":"Time-series forecasting with deep learning: A survey","volume":"379","author":"Lim Bryan","year":"2021","unstructured":"Bryan Lim and Stefan Zohren. 2021. Time-series forecasting with deep learning: A survey. Philosophical Transactions of the Royal Society A 379, 2194 (2021), 20200209.","journal-title":"Philosophical Transactions of the Royal Society A"},{"key":"e_1_3_2_16_2","doi-asserted-by":"publisher","unstructured":"Jiawei Liu Chunqiu Steven Xia Yuyao Wang and Lingming Zhang. 2023. Is your code generated by ChatGPT really correct? Rigorous evaluation of large language models for code generation. DOI: 10.48550\/arXiv.2305.01210","DOI":"10.48550\/arXiv.2305.01210"},{"key":"e_1_3_2_17_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-19034-7_14"},{"key":"e_1_3_2_18_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10462-022-10246-w"},{"key":"e_1_3_2_19_2","doi-asserted-by":"publisher","DOI":"10.4249\/scholarpedia.1883"},{"issue":"8","key":"e_1_3_2_20_2","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford Alec","year":"2019","unstructured":"Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI Blog 1, 8 (2019), 9.","journal-title":"OpenAI Blog"},{"key":"e_1_3_2_21_2","doi-asserted-by":"crossref","unstructured":"Hassan Ramchoun Youssef Ghanou Mohamed Ettaouil and Mohammed Amine Janati Idrissi. 2016. Multilayer perceptron: Architecture optimization and training.","DOI":"10.9781\/ijimai.2016.415"},{"key":"e_1_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4842-9529-8"},{"key":"e_1_3_2_23_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4842-9529-8_5"},{"key":"e_1_3_2_24_2","doi-asserted-by":"publisher","DOI":"10.1145\/3491101.3519665"},{"key":"e_1_3_2_25_2","doi-asserted-by":"publisher","unstructured":"Jules White Sam Hays Quchen Fu Jesse Spencer-Smith and Douglas C. Schmidt. 2023. Chatgpt prompt patterns for improving code quality refactoring requirements elicitation and software design. DOI: 10.48550\/arXiv.2303.07839","DOI":"10.48550\/arXiv.2303.07839"},{"key":"e_1_3_2_26_2","doi-asserted-by":"publisher","DOI":"10.1145\/3558489.3559072"}],"container-title":["ACM Transactions on Intelligent Systems and Technology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3663485","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,24]],"date-time":"2025-11-24T15:12:38Z","timestamp":1763997158000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3663485"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,24]]},"references-count":25,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2025,12,31]]}},"alternative-id":["10.1145\/3663485"],"URL":"https:\/\/doi.org\/10.1145\/3663485","relation":{},"ISSN":["2157-6904","2157-6912"],"issn-type":[{"value":"2157-6904","type":"print"},{"value":"2157-6912","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,11,24]]},"assertion":[{"value":"2023-12-04","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-04-26","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-11-24","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}