Python 每日数据刮取

Python 每日数据刮取,python,web-scraping,Python,Web Scraping,我正在尝试使用Python在未来30天内每天计算同一家公司的股价。我使用了列表索引和.append(),一旦添加了更新的价格,初始值就会被替换。我怎样才能列出30天内相同股票的价格 *#Catalyst Pharmaceuticals #New York Stack Exchange import requests import pytz from bs4 import BeautifulSoup import datetime import csv r=requests.get('https:

我正在尝试使用Python在未来30天内每天计算同一家公司的股价。我使用了列表索引和
.append()
,一旦添加了更新的价格,初始值就会被替换。我怎样才能列出30天内相同股票的价格

*#Catalyst Pharmaceuticals
#New York Stack Exchange
import requests
import pytz
from bs4 import BeautifulSoup
import datetime
import csv
r=requests.get('https://robinhood.com/collections/technology')
html=r.content
soup=BeautifulSoup(html,'html.parser')
csv_file=open('Catalyst Pharmaceuticals Monthly.csv','a')
csv_writer=csv.writer(csv_file)
price_list = []
dttm = []
def websc():
    
    global price_list
    global dttm
    global a_price
    #i=10
        
    
    for p in soup.find_all('a',{'class':'rh-hyperlink'})[2]:
        a_price = p.text
        dd=datetime.datetime.now(pytz.timezone("GMT"))
        dd=dd.strftime("%Y-%m-%d %H:%M:%S")
        price_list.append(a_price)
        dttm.append(dd)
        
       
      
    
    zipped = zip(price_list,dttm)
    d = list(zipped)
    print(d)
    csv_writer.writerows(d)
    csv_file.close()
websc()*

如果您不想覆盖文件,您需要以追加模式而不是写入模式打开文件。难道您不能通过一些标记器循环,将所有内容推送到数据框中,然后将其导出到CSV吗

import pandas as pd  
import numpy as np
import matplotlib.pyplot as plt
import seaborn as sns
import scipy.optimize as sco
import datetime as dt
import math
from datetime import datetime, timedelta
from pandas_datareader import data as wb
from sklearn.cluster import KMeans
np.random.seed(777)


start = '2020-01-01'
end = '2020-08-27'
#N = 165
#start = datetime.now() - timedelta(days=N)
#end = dt.datetime.today()



tickers = ['AAPL','MSFT','GOOG','SBUX','MCD','NKE']

thelen = len(tickers)

price_data = []
for ticker in tickers:
    try:
        prices = wb.DataReader(ticker, start = start, end = end, data_source='yahoo')[['Adj Close']]
        price_data.append(prices.assign(ticker=ticker)[['ticker', 'Adj Close']])
    except:
        print(ticker)
            
df = pd.concat(price_data)
df.dtypes
df.head()
df.shape


# finally....
df.to_csv('file_name.csv')

如果您需要与此相关的其他内容,请尝试此操作并发回。

首先:如果您仅将.append()添加到列表中,则永远不会替换数据。第二,为什么你要用机器学习和深度学习来标记它第二个条目比第一个条目低5行。你是在文件中添加空行吗?