from bs4 import BeautifulSoup from selenium import webdriver import pandas as pd import csv import time #作成したExcel/Csvを保存するフォルダを指定 export_file_path="**********" #ブラウザの立ち上げ browser=webdriver.Chrome("**********") browser.implicitly_wait(1) url="**********" #指定したURLへ遷移する browser.get(url) #検索ボックスを指定 elem=browser.find_element_by_name("**********") elem.clear() #任意のキーワードをinput()関数にて入力 elem.send_keys(inp
![任意の単語で検索しタイトルとURLを抽出するスクレイピング - Qiita](https://cdn-ak-scissors.b.st-hatena.com/image/square/7f7a2736fdd30886df4d0c5d62fb35dbaa991464/height=288;version=1;width=512/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fcdn.qiita.com%252Fassets%252Fpublic%252Farticle-ogp-background-9f5428127621718a910c8b63951390ad.png%3Fixlib%3Drb-4.0.0%26w%3D1200%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTkxNiZ0eHQ9JUU0JUJCJUJCJUU2JTg0JThGJUUzJTgxJUFFJUU1JThEJTk4JUU4JUFBJTlFJUUzJTgxJUE3JUU2JUE0JTlDJUU3JUI0JUEyJUUzJTgxJTk3JUUzJTgyJUJGJUUzJTgyJUE0JUUzJTgzJTg4JUUzJTgzJUFCJUUzJTgxJUE4VVJMJUUzJTgyJTkyJUU2JThBJUJEJUU1JTg3JUJBJUUzJTgxJTk5JUUzJTgyJThCJUUzJTgyJUI5JUUzJTgyJUFGJUUzJTgzJUFDJUUzJTgyJUE0JUUzJTgzJTk0JUUzJTgzJUIzJUUzJTgyJUIwJnR4dC1jb2xvcj0lMjMyMTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9NTYmdHh0LWNsaXA9ZWxsaXBzaXMmdHh0LWFsaWduPWxlZnQlMkN0b3Amcz1iNmE0MzQ3ZjQ0ZjNmNjBkNjU5MWQ2NGY0ODk3N2Y3OA%26mark-x%3D142%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTYxNiZ0eHQ9JTQwdGFrZXNoaXcwMiZ0eHQtY29sb3I9JTIzMjEyMTIxJnR4dC1mb250PUhpcmFnaW5vJTIwU2FucyUyMFc2JnR4dC1zaXplPTM2JnR4dC1hbGlnbj1sZWZ0JTJDdG9wJnM9Y2RiMjNkZDY2ZWIxZTFiMTZmMjhlYmNhYzRlOTJhNjk%26blend-x%3D142%26blend-y%3D491%26blend-mode%3Dnormal%26s%3Ddd2bcbf38a0051fa7d8bf17139a47eec)