Enhancing Zero-Shot Crypto Sentiment with Fine-tuned Language Model and Prompt Engineering (2310.13226v1)
Abstract: Blockchain technology has revolutionized the financial landscape, with cryptocurrencies gaining widespread adoption for their decentralized and transparent nature. As the sentiment expressed on social media platforms can significantly influence cryptocurrency discussions and market movements, sentiment analysis has emerged as a crucial tool for understanding public opinion and predicting market trends. Motivated by the aim to enhance sentiment analysis accuracy in the cryptocurrency domain, this paper investigates fine-tuning techniques on LLMs. This paper also investigates the efficacy of supervised fine-tuning and instruction-based fine-tuning on LLMs for unseen tasks. Experimental results demonstrate a significant average zero-shot performance gain of 40% after fine-tuning, highlighting the potential of this technique in optimizing pre-trained LLM efficiency. Additionally, the impact of instruction tuning on models of varying scales is examined, revealing that larger models benefit from instruction tuning, achieving the highest average accuracy score of 75.16%. In contrast, smaller-scale models may experience reduced generalization due to the complete utilization of model capacity. To gain deeper insight about how instruction works with these LLMs, this paper presents an experimental investigation into the response of an instruction-based model under different instruction tuning setups. The investigation demonstrates that the model achieves an average accuracy score of 72.38% for short and simple instructions. This performance significantly outperforms its accuracy under long and complex instructions by over 12%, thereby effectively highlighting the profound significance of instruction characteristics in maximizing model performance.
- Towards characterizing blockchain-based cryptocurrencies for highly-accurate predictions. INFOCOM 2018 - IEEE Conference on Computer Communications Workshops, pages 704–709, 7 2018.
- Vitalik Buterin. Ethereum: A next-generation smart contract and decentralized application platform. 2014.
- Improving the cryptocurrency price prediction performance based on reinforcement learning. IEEE Access, 9:162651–162659, 2021.
- Cryptocurrency trading: a comprehensive survey. Financial Innovation, 8:1–59, 12 2022.
- Stochastic neural networks for cryptocurrency price prediction. IEEE Access, 8:82804–82818, 2020.
- Tweet sentiment analysis for cryptocurrencies. Proceedings - 6th International Conference on Computer Science and Engineering, UBMK 2021, pages 613–618, 2021.
- Pratikkumar Prajapati. Predictive analysis of bitcoin price considering social sentiments. 1 2020.
- Multisource financial sentiment analysis for detecting bitcoin price change indications using deep learning. Neural Computing and Applications, 34:19441–19452, 11 2022.
- Using sentiment analysis to predict opinion inversion in tweets of political communication. Scientific Reports 2021 11:1, 11:1–9, 3 2021.
- Deep learning based sentiment analysis and offensive language identification on multilingual code-mixed data. Scientific Reports 2022 12:1, 12:1–12, 12 2022.
- Binder Tweet. Twitter real time monitoring and twitter live analytics. https://www.tweetbinder.com/blog/real-time-twitter/, 2019.
- Transformer-based deep learning models for the sentiment analysis of social media data. Array, 14:100157, 7 2022.
- A comparative analysis of recommender systems based on item aspect opinions extracted from user reviews. User Modeling and User-Adapted Interaction, 29:381–441, 4 2019.
- A critical review on sentiment analysis techniques. Proceedings of 3rd International Conference on Intelligent Engineering and Management, ICIEM 2022, pages 741–746, 2022.
- Deep convolution neural networks for twitter sentiment analysis. IEEE Access, 6:23253–23260, 1 2018.
- A comprehensive analysis of adverb types for mining user sentiments on amazon product reviews. World Wide Web, 23:1811–1829, 5 2020.
- Evaluation of deep learning techniques in sentiment analysis from twitter data. Proceedings - 2019 International Conference on Deep Learning and Machine Learning in Emerging Applications, Deep-ML 2019, pages 12–17, 8 2019.
- Glove: Global vectors for word representation. pages 1532–1543. Association for Computational Linguistics, 2014.
- Attention is all you need. Advances in Neural Information Processing Systems, 2017-December:5999–6009, 6 2017.
- Bert: Pre-training of deep bidirectional transformers for language understanding. 10 2018.
- Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 21:1–67, 10 2019.
- Using deepspeed and megatron to train megatron-turing nlg 530b, a large-scale generative language model. 1 2022.
- Llama: Open and efficient foundation language models. 2 2023.
- Finetuned language models are zero-shot learners. 9 2021.
- A survey of sentiment analysis in social media. Knowledge and Information Systems, 60:617–663, 8 2019.
- Sentiment analysis on cryptocurrency based on tweets and retweets using support vector machines and chi-square. 2022 7th International Conference on Informatics and Computing, ICIC 2022, 2022.
- Cryptocurrency sentiment analysis on the twitter platform using support vector machine (svm) algorithm. Proceedings - International Conference Advancement in Data Science, E-Learning and Information Systems, ICADEIS 2022, 2022.
- Sentiment analysis of bitcoin data by tweets through naive bayes. pages 1–4. IEEE, 11 2022.
- Sentiment analysis on cryptocurrency using youtube comments. pages 730–733. IEEE, 3 2022.
- On the predictive power of tweet sentiments and attention on bitcoin. International Review of Economics & Finance, 79:289–301, 5 2022.
- Cryptocurrency price prediction using social media sentiment analysis. 13th International Conference on Information, Intelligence, Systems and Applications, IISA 2022, 2022.
- Twitter sentiment analysis for bitcoin price prediction. 2023 3rd International Conference on Smart Data Intelligence (ICSMDI), pages 32–37, 3 2023.
- ¡i¿dl-guess¡/i¿ : Deep learning and sentiment analysis-based cryptocurrency price prediction. IEEE Access, 10:35398–35409, 2022.
- Himanshu Dwivedi. Cryptocurrency sentiment analysis using bidirectional transformation. pages 140–142. IEEE, 3 2023.
- Cbits: Crypto bert incorporated trading system. IEEE Access, 11:6912–6921, 2023.
- Sentiment analysis towards cryptocurrency and nft in bahasa indonesia for twitter large amount data using bert. International Journal of Intelligent Systems and Applications in Engineering, 11:303 – 309, 2 2023.
- On technical trading and social media indicators in cryptocurrencies’ price classification through deep learning. 2 2021.
- Pre-trained models for natural language processing: A survey. 3 2020.
- Palm: Scaling language modeling with pathways. 4 2022.
- Training compute-optimal large language models. 3 2022.
- The refinedweb dataset for falcon llm: Outperforming curated corpora with web data, and web data only. 6 2023.
- Enabling ai-generated content (aigc) services in wireless edge networks. 1 2023.
- A brief overview of chatgpt: The history, status quo and potential future development. IEEE/CAA Journal of Automatica Sinica, 10:1122–1136, 5 2023.
- Language models are few-shot learners. Advances in Neural Information Processing Systems, 33:1877–1901, 2020.
- Opt: Open pre-trained transformer language models. 5 2022.
- Training language models to follow instructions with human feedback. 3 2022.
- A survey of large language models. 3 2023.
- An overview on language models: Recent developments and outlook. 3 2023.
- Raschka Sebastian. Finetuning large language models. https://magazine.sebastianraschka.com/p/finetuning-large-language-models?subscribe_prompt=free, 4 2023.
- Shah Kushal. Pre-training, fine-tuning and in-context learning in large language models (llms) — by kushal shah — medium. https://medium.com/@atmabodha/pre-training-fine-tuning-and-in-context-learning-in-large-language-models-llms-dd483707b122, 8 2022.
- Spdf: Sparse pre-training and dense fine-tuning for large language models. 3 2023.
- Multitask prompted training enables zero-shot task generalization. 10 2021.
- Scaling instruction-finetuned language models. 10 2022.
- Crosslingual generalization through multitask finetuning. 11 2022.
- Bloom: A 176b-parameter open-access multilingual language model. 11 2022.
- A survey on in-context learning. 12 2022.
- Layton Dennis. Prompt engineering - leveraging in context learning. https://www.linkedin.com/pulse/prompt-engineering-leveraging-context-learning-dennis-layton/, 4 2023.
- C. Montgomery and Patrick H. Winston. Learning and reasoning by analogy. Communications of the ACM, 23:689–703, 12 1980.
- Metaicl: Learning to learn in context. NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference, pages 2791–2809, 10 2021.
- Why can gpt learn in-context? language models implicitly perform gradient descent as meta-optimizers. 12 2022.
- Can chatgpt forecast stock price movements? return predictability and large language models. 4 2023.
- A prompt pattern catalog to enhance prompt engineering with chatgpt. 2 2023.
- Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Computing Surveys, 55, 7 2021.
- Making pre-trained language models better few-shot learners. pages 3816–3830. Association for Computational Linguistics, 2021.
- How many data points is a prompt worth? NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference, pages 2627–2636, 3 2021.
- It’s not just size that matters: Small language models are also few-shot learners. NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference, pages 2339–2352, 9 2020.
- Interactive and visual prompt engineering for ad-hoc task adaptation with large language models. IEEE Transactions on Visualization and Computer Graphics, 29:1146–1156, 1 2023.
- Chat2vis: Generating data visualisations via natural language using chatgpt, codex and gpt-3 large language models. 2 2023.
- OpenAI Platform. Gpt best practices - openai api. https://platform.openai.com/docs/guides/gpt-best-practices.