{"id":36089,"date":"2024-11-01T09:45:41","date_gmt":"2024-11-01T09:45:41","guid":{"rendered":"http:\/\/atmokpo.com\/w\/?p=36089"},"modified":"2024-11-01T09:45:41","modified_gmt":"2024-11-01T09:45:41","slug":"hugging-face-transformers-course-setting-up-the-bigbird-library-and-loading-pre-trained-models","status":"publish","type":"post","link":"https:\/\/atmokpo.com\/w\/36089\/","title":{"rendered":"Hugging Face Transformers Course, Setting Up the BigBird Library and Loading Pre-trained Models"},"content":{"rendered":"<p><body><\/p>\n<p>Recently, transformer-based models have been gaining attention in the field of Natural Language Processing (NLP) due to their outstanding performance. Among them, <strong>BigBird<\/strong>, developed by Google, is an innovative architecture designed for large-scale document understanding and processing long sequences. In this course, we will learn how to set up the BigBird model using Hugging Face&#8217;s transformers library and how to load a pre-trained model.<\/p>\n<h2>1. What is BigBird?<\/h2>\n<p>BigBird is a model designed as an extension of the Transformer model, particularly created to efficiently handle long sequence data. Traditional Transformer models have limitations on the length of the input sequence, usually processing only up to about 512 tokens of text. In contrast, BigBird overcomes this limitation using a sparse attention mechanism. This is useful for various NLP tasks such as document summarization, question answering, and text classification.<\/p>\n<h3>1.1 Key Features of BigBird<\/h3>\n<ul>\n<li>Ability to process long input sequences<\/li>\n<li>Reduces memory consumption and improves processing speed<\/li>\n<li>Easy to apply to various NLP tasks by utilizing pre-trained models<\/li>\n<\/ul>\n<h2>2. Setting Up the Environment<\/h2>\n<p>To use the BigBird model, you need to set up your Python environment. Follow the steps below to proceed with the installation.<\/p>\n<h3>2.1 Installing Python and pip<\/h3>\n<p>You need Python version 3.6 or higher. You can install Python and pip with the following commands:<\/p>\n<pre><code>sudo apt update\nsudo apt install python3 python3-pip<\/code><\/pre>\n<h3>2.2 Installing Hugging Face Transformers Library<\/h3>\n<p>Use the command below to install Hugging Face&#8217;s transformers library:<\/p>\n<pre><code>pip install transformers<\/code><\/pre>\n<h3>2.3 Installing Additional Libraries<\/h3>\n<p>Additional libraries also need to be installed to use the BigBird model:<\/p>\n<pre><code>pip install torch<\/code><\/pre>\n<h2>3. Loading the Pre-Trained Model<\/h2>\n<p>Now that all the settings are complete, we are ready to load and use the BigBird model. We will use Hugging Face&#8217;s <code>transformers<\/code> library for this.<\/p>\n<h3>3.1 Text Summarization<\/h3>\n<p>Let&#8217;s take a look at an example of text summarization using the BigBird model. Refer to the code below:<\/p>\n<pre><code>from transformers import BigBirdTokenizer, BigBirdForSequenceClassification\n\n# Load the tokenizer and model\ntokenizer = BigBirdTokenizer.from_pretrained('google\/bigbird-roberta-base')\nmodel = BigBirdForSequenceClassification.from_pretrained('google\/bigbird-roberta-base')\n\n# Input text\ntext = \"Deep learning is a branch of machine learning that utilizes artificial neural networks. It is used to learn patterns from data and make predictions and decisions based on this.\"\n\n# Tokenize the text and convert it to tensor\ninputs = tokenizer(text, return_tensors=\"pt\", padding=True, truncation=True)\n\n# Model prediction\noutputs = model(**inputs)\nlogits = outputs.logits\npredicted_class = logits.argmax().item()\n\nprint(f\"Predicted class: {predicted_class}\")<\/code><\/pre>\n<h4>Code Explanation<\/h4>\n<p>In the code above, we use the <code>BigBirdTokenizer<\/code> and <code>BigBirdForSequenceClassification<\/code> classes to load the pre-trained BigBird model and tokenizer.<\/p>\n<ul>\n<li>We load Google\u2019s pre-trained BigBird model using the <code>from_pretrained<\/code> method.<\/li>\n<li>To tokenize the input text, we use <code>tokenizer<\/code> to convert the text into a tensor.<\/li>\n<li>To check the model&#8217;s prediction results, we perform an argmax operation on the output logits to predict the class.<\/li>\n<\/ul>\n<h3>3.2 Training the Model<\/h3>\n<p>Now, let&#8217;s look at how to further train the pre-trained model on a specific dataset. Below is a code showing a simple training routine:<\/p>\n<pre><code>from transformers import Trainer, TrainingArguments\nfrom datasets import load_dataset\n\n# Load the dataset (e.g., IMDB sentiment analysis dataset)\ndataset = load_dataset('imdb')\n\n# Set training arguments\ntraining_args = TrainingArguments(\n    output_dir='.\/results',\n    num_train_epochs=1,\n    per_device_train_batch_size=8,\n    save_steps=10_000,\n    save_total_limit=2,\n)\n\n# Create a Trainer object\ntrainer = Trainer(\n    model=model,\n    args=training_args,\n    train_dataset=dataset['train'],\n    eval_dataset=dataset['test'],\n)\n\n# Train the model\ntrainer.train()<\/code><\/pre>\n<h4>Code Explanation<\/h4>\n<p>In the code above, we load the IMDB sentiment analysis dataset using the <code>datasets<\/code> library. We perform training on the BigBird model based on this dataset:<\/p>\n<ul>\n<li>We specify various training settings (epochs, batch size, etc.) using <code>TrainingArguments<\/code>.<\/li>\n<li>The <code>Trainer<\/code> class allows us to perform training and evaluation.<\/li>\n<\/ul>\n<h2>4. Summary<\/h2>\n<p>In this course, we learned how to set up the BigBird model using the Hugging Face transformers library and how to load a pre-trained model. BigBird is a powerful tool that can efficiently process long input sequences. By applying it to various NLP tasks, we can significantly enhance performance, and we can optimize the model through fine-tuning for specific tasks.<\/p>\n<p>We hope you continue exploring how to utilize models like BigBird in various deep learning projects. If you need additional materials or have questions, please leave a comment! Thank you.<\/p>\n<p><\/body><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Recently, transformer-based models have been gaining attention in the field of Natural Language Processing (NLP) due to their outstanding performance. Among them, BigBird, developed by Google, is an innovative architecture designed for large-scale document understanding and processing long sequences. In this course, we will learn how to set up the BigBird model using Hugging Face&#8217;s &hellip; <a href=\"https:\/\/atmokpo.com\/w\/36089\/\" class=\"more-link\">\ub354 \ubcf4\uae30<span class=\"screen-reader-text\"> &#8220;Hugging Face Transformers Course, Setting Up the BigBird Library and Loading Pre-trained Models&#8221;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_jetpack_memberships_contains_paid_content":false,"footnotes":""},"categories":[108],"tags":[],"class_list":["post-36089","post","type-post","status-publish","format-standard","hentry","category---en"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v26.2 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Hugging Face Transformers Course, Setting Up the BigBird Library and Loading Pre-trained Models - \ub77c\uc774\ube0c\uc2a4\ub9c8\ud2b8<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/atmokpo.com\/w\/36089\/\" \/>\n<meta property=\"og:locale\" content=\"ko_KR\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Hugging Face Transformers Course, Setting Up the BigBird Library and Loading Pre-trained Models - \ub77c\uc774\ube0c\uc2a4\ub9c8\ud2b8\" \/>\n<meta property=\"og:description\" content=\"Recently, transformer-based models have been gaining attention in the field of Natural Language Processing (NLP) due to their outstanding performance. Among them, BigBird, developed by Google, is an innovative architecture designed for large-scale document understanding and processing long sequences. In this course, we will learn how to set up the BigBird model using Hugging Face&#8217;s &hellip; \ub354 \ubcf4\uae30 &quot;Hugging Face Transformers Course, Setting Up the BigBird Library and Loading Pre-trained Models&quot;\" \/>\n<meta property=\"og:url\" content=\"https:\/\/atmokpo.com\/w\/36089\/\" \/>\n<meta property=\"og:site_name\" content=\"\ub77c\uc774\ube0c\uc2a4\ub9c8\ud2b8\" \/>\n<meta property=\"article:published_time\" content=\"2024-11-01T09:45:41+00:00\" \/>\n<meta name=\"author\" content=\"root\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@bebubo4\" \/>\n<meta name=\"twitter:site\" content=\"@bebubo4\" \/>\n<meta name=\"twitter:label1\" content=\"\uae00\uc4f4\uc774\" \/>\n\t<meta name=\"twitter:data1\" content=\"root\" \/>\n\t<meta name=\"twitter:label2\" content=\"\uc608\uc0c1 \ub418\ub294 \ud310\ub3c5 \uc2dc\uac04\" \/>\n\t<meta name=\"twitter:data2\" content=\"4\ubd84\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/atmokpo.com\/w\/36089\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/atmokpo.com\/w\/36089\/\"},\"author\":{\"name\":\"root\",\"@id\":\"https:\/\/atmokpo.com\/w\/#\/schema\/person\/91b6b3b138fbba0efb4ae64b1abd81d7\"},\"headline\":\"Hugging Face Transformers Course, Setting Up the BigBird Library and Loading Pre-trained Models\",\"datePublished\":\"2024-11-01T09:45:41+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/atmokpo.com\/w\/36089\/\"},\"wordCount\":534,\"publisher\":{\"@id\":\"https:\/\/atmokpo.com\/w\/#organization\"},\"articleSection\":[\"Using Hugging Face\"],\"inLanguage\":\"ko-KR\"},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/atmokpo.com\/w\/36089\/\",\"url\":\"https:\/\/atmokpo.com\/w\/36089\/\",\"name\":\"Hugging Face Transformers Course, Setting Up the BigBird Library and Loading Pre-trained Models - \ub77c\uc774\ube0c\uc2a4\ub9c8\ud2b8\",\"isPartOf\":{\"@id\":\"https:\/\/atmokpo.com\/w\/#website\"},\"datePublished\":\"2024-11-01T09:45:41+00:00\",\"breadcrumb\":{\"@id\":\"https:\/\/atmokpo.com\/w\/36089\/#breadcrumb\"},\"inLanguage\":\"ko-KR\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/atmokpo.com\/w\/36089\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/atmokpo.com\/w\/36089\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"\ud648\",\"item\":\"https:\/\/atmokpo.com\/w\/en\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Hugging Face Transformers Course, Setting Up the BigBird Library and Loading Pre-trained Models\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/atmokpo.com\/w\/#website\",\"url\":\"https:\/\/atmokpo.com\/w\/\",\"name\":\"\ub77c\uc774\ube0c\uc2a4\ub9c8\ud2b8\",\"description\":\"\",\"publisher\":{\"@id\":\"https:\/\/atmokpo.com\/w\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/atmokpo.com\/w\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"ko-KR\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/atmokpo.com\/w\/#organization\",\"name\":\"\ub77c\uc774\ube0c\uc2a4\ub9c8\ud2b8\",\"url\":\"https:\/\/atmokpo.com\/w\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"ko-KR\",\"@id\":\"https:\/\/atmokpo.com\/w\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/atmokpo.com\/w\/wp-content\/uploads\/2024\/11\/logo.png\",\"contentUrl\":\"https:\/\/atmokpo.com\/w\/wp-content\/uploads\/2024\/11\/logo.png\",\"width\":400,\"height\":400,\"caption\":\"\ub77c\uc774\ube0c\uc2a4\ub9c8\ud2b8\"},\"image\":{\"@id\":\"https:\/\/atmokpo.com\/w\/#\/schema\/logo\/image\/\"},\"sameAs\":[\"https:\/\/x.com\/bebubo4\"]},{\"@type\":\"Person\",\"@id\":\"https:\/\/atmokpo.com\/w\/#\/schema\/person\/91b6b3b138fbba0efb4ae64b1abd81d7\",\"name\":\"root\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"ko-KR\",\"@id\":\"https:\/\/atmokpo.com\/w\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/708197b41fc6435a7ce22d951b25d4a47e9e904270cb1f04682d4f025066f80c?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/708197b41fc6435a7ce22d951b25d4a47e9e904270cb1f04682d4f025066f80c?s=96&d=mm&r=g\",\"caption\":\"root\"},\"sameAs\":[\"http:\/\/atmokpo.com\/w\"],\"url\":\"https:\/\/atmokpo.com\/w\/author\/root\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Hugging Face Transformers Course, Setting Up the BigBird Library and Loading Pre-trained Models - \ub77c\uc774\ube0c\uc2a4\ub9c8\ud2b8","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/atmokpo.com\/w\/36089\/","og_locale":"ko_KR","og_type":"article","og_title":"Hugging Face Transformers Course, Setting Up the BigBird Library and Loading Pre-trained Models - \ub77c\uc774\ube0c\uc2a4\ub9c8\ud2b8","og_description":"Recently, transformer-based models have been gaining attention in the field of Natural Language Processing (NLP) due to their outstanding performance. Among them, BigBird, developed by Google, is an innovative architecture designed for large-scale document understanding and processing long sequences. In this course, we will learn how to set up the BigBird model using Hugging Face&#8217;s &hellip; \ub354 \ubcf4\uae30 \"Hugging Face Transformers Course, Setting Up the BigBird Library and Loading Pre-trained Models\"","og_url":"https:\/\/atmokpo.com\/w\/36089\/","og_site_name":"\ub77c\uc774\ube0c\uc2a4\ub9c8\ud2b8","article_published_time":"2024-11-01T09:45:41+00:00","author":"root","twitter_card":"summary_large_image","twitter_creator":"@bebubo4","twitter_site":"@bebubo4","twitter_misc":{"\uae00\uc4f4\uc774":"root","\uc608\uc0c1 \ub418\ub294 \ud310\ub3c5 \uc2dc\uac04":"4\ubd84"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/atmokpo.com\/w\/36089\/#article","isPartOf":{"@id":"https:\/\/atmokpo.com\/w\/36089\/"},"author":{"name":"root","@id":"https:\/\/atmokpo.com\/w\/#\/schema\/person\/91b6b3b138fbba0efb4ae64b1abd81d7"},"headline":"Hugging Face Transformers Course, Setting Up the BigBird Library and Loading Pre-trained Models","datePublished":"2024-11-01T09:45:41+00:00","mainEntityOfPage":{"@id":"https:\/\/atmokpo.com\/w\/36089\/"},"wordCount":534,"publisher":{"@id":"https:\/\/atmokpo.com\/w\/#organization"},"articleSection":["Using Hugging Face"],"inLanguage":"ko-KR"},{"@type":"WebPage","@id":"https:\/\/atmokpo.com\/w\/36089\/","url":"https:\/\/atmokpo.com\/w\/36089\/","name":"Hugging Face Transformers Course, Setting Up the BigBird Library and Loading Pre-trained Models - \ub77c\uc774\ube0c\uc2a4\ub9c8\ud2b8","isPartOf":{"@id":"https:\/\/atmokpo.com\/w\/#website"},"datePublished":"2024-11-01T09:45:41+00:00","breadcrumb":{"@id":"https:\/\/atmokpo.com\/w\/36089\/#breadcrumb"},"inLanguage":"ko-KR","potentialAction":[{"@type":"ReadAction","target":["https:\/\/atmokpo.com\/w\/36089\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/atmokpo.com\/w\/36089\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"\ud648","item":"https:\/\/atmokpo.com\/w\/en\/"},{"@type":"ListItem","position":2,"name":"Hugging Face Transformers Course, Setting Up the BigBird Library and Loading Pre-trained Models"}]},{"@type":"WebSite","@id":"https:\/\/atmokpo.com\/w\/#website","url":"https:\/\/atmokpo.com\/w\/","name":"\ub77c\uc774\ube0c\uc2a4\ub9c8\ud2b8","description":"","publisher":{"@id":"https:\/\/atmokpo.com\/w\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/atmokpo.com\/w\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"ko-KR"},{"@type":"Organization","@id":"https:\/\/atmokpo.com\/w\/#organization","name":"\ub77c\uc774\ube0c\uc2a4\ub9c8\ud2b8","url":"https:\/\/atmokpo.com\/w\/","logo":{"@type":"ImageObject","inLanguage":"ko-KR","@id":"https:\/\/atmokpo.com\/w\/#\/schema\/logo\/image\/","url":"https:\/\/atmokpo.com\/w\/wp-content\/uploads\/2024\/11\/logo.png","contentUrl":"https:\/\/atmokpo.com\/w\/wp-content\/uploads\/2024\/11\/logo.png","width":400,"height":400,"caption":"\ub77c\uc774\ube0c\uc2a4\ub9c8\ud2b8"},"image":{"@id":"https:\/\/atmokpo.com\/w\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/x.com\/bebubo4"]},{"@type":"Person","@id":"https:\/\/atmokpo.com\/w\/#\/schema\/person\/91b6b3b138fbba0efb4ae64b1abd81d7","name":"root","image":{"@type":"ImageObject","inLanguage":"ko-KR","@id":"https:\/\/atmokpo.com\/w\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/708197b41fc6435a7ce22d951b25d4a47e9e904270cb1f04682d4f025066f80c?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/708197b41fc6435a7ce22d951b25d4a47e9e904270cb1f04682d4f025066f80c?s=96&d=mm&r=g","caption":"root"},"sameAs":["http:\/\/atmokpo.com\/w"],"url":"https:\/\/atmokpo.com\/w\/author\/root\/"}]}},"jetpack_featured_media_url":"","jetpack_sharing_enabled":true,"jetpack-related-posts":[],"_links":{"self":[{"href":"https:\/\/atmokpo.com\/w\/wp-json\/wp\/v2\/posts\/36089","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/atmokpo.com\/w\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/atmokpo.com\/w\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/atmokpo.com\/w\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/atmokpo.com\/w\/wp-json\/wp\/v2\/comments?post=36089"}],"version-history":[{"count":1,"href":"https:\/\/atmokpo.com\/w\/wp-json\/wp\/v2\/posts\/36089\/revisions"}],"predecessor-version":[{"id":36090,"href":"https:\/\/atmokpo.com\/w\/wp-json\/wp\/v2\/posts\/36089\/revisions\/36090"}],"wp:attachment":[{"href":"https:\/\/atmokpo.com\/w\/wp-json\/wp\/v2\/media?parent=36089"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/atmokpo.com\/w\/wp-json\/wp\/v2\/categories?post=36089"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/atmokpo.com\/w\/wp-json\/wp\/v2\/tags?post=36089"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}