HOME 首页
SERVICE 服务产品
XINMEITI 新媒体代运营
CASE 服务案例
NEWS 热点资讯
ABOUT 关于我们
CONTACT 联系我们
创意岭
让品牌有温度、有情感
专注品牌策划15年

    openai免费使用(open aip)

    发布时间:2023-03-12 18:02:04     稿源: 创意岭    阅读: 77        问大家

    大家好!今天让创意岭的小编来大家介绍下关于openai免费使用的问题,以下是小编对此问题的归纳整理,让我们一起来看看吧。

    ChatGPT国内免费在线使用,一键生成原创文章、方案、文案、工作计划、工作报告、论文、代码、作文、做题和对话答疑等等

    只需要输入关键词,就能返回你想要的内容,越精准,写出的就越详细,有微信小程序端、在线网页版、PC客户端

    官网:https://ai.de1919.com

    本文目录:

    openai免费使用(open aip)

    一、openai能当爬虫使吗

    你好,可以的,Spinning Up是OpenAI开源的面向初学者的深度强化学习资料,其中列出了105篇深度强化学习领域非常经典的文章, 见 Spinning Up:

    博主使用Python爬虫自动爬取了所有文章,而且爬下来的文章也按照网页的分类自动分类好。

    见下载资源:Spinning Up Key Papers

    源码如下:

    import os

    import time

    import urllib.request as url_re

    import requests as rq

    from bs4 import BeautifulSoup as bf

    '''Automatically download all the key papers recommended by OpenAI Spinning Up.

    See more info on: https://spinningup.openai.com/en/latest/spinningup/keypapers.html

    Dependency:

    bs4, lxml

    '''

    headers = {

    'User-Agent':'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/92.0.4515.131 Safari/537.36'

    }

    spinningup_url = 'https://spinningup.openai.com/en/latest/spinningup/keypapers.html'

    paper_id = 1

    def download_pdf(pdf_url, pdf_path):

    """Automatically download PDF file from Internet

    Args:

    pdf_url (str): url of the PDF file to be downloaded

    pdf_path (str): save routine of the downloaded PDF file

    """

    if os.path.exists(pdf_path): return

    try:

    with url_re.urlopen(pdf_url) as url:

    pdf_data = url.read()

    with open(pdf_path, "wb") as f:

    f.write(pdf_data)

    except: # fix link at [102]

    pdf_url = r"https://is.tuebingen.mpg.de/fileadmin/user_upload/files/publications/Neural-Netw-2008-21-682_4867%5b0%5d.pdf"

    with url_re.urlopen(pdf_url) as url:

    pdf_data = url.read()

    with open(pdf_path, "wb") as f:

    f.write(pdf_data)

    time.sleep(10) # sleep 10 seconds to download next

    def download_from_bs4(papers, category_path):

    """Download papers from Spinning Up

    Args:

    papers (bs4.element.ResultSet): 'a' tags with paper link

    category_path (str): root dir of the paper to be downloaded

    """

    global paper_id

    print("Start to ownload papers from catagory {}...".format(category_path))

    for paper in papers:

    paper_link = paper['href']

    if not paper_link.endswith('.pdf'):

    if paper_link[8:13] == 'arxiv':

    # paper_link = "https://arxiv.org/abs/1811.02553"

    paper_link = paper_link[:18] + 'pdf' + paper_link[21:] + '.pdf' # arxiv link

    elif paper_link[8:18] == 'openreview': # openreview link

    # paper_link = "https://openreview.net/forum?id=ByG_3s09KX"

    paper_link = paper_link[:23] + 'pdf' + paper_link[28:]

    elif paper_link[14:18] == 'nips': # neurips link

    paper_link = "https://proceedings.neurips.cc/paper/2017/file/a1d7311f2a312426d710e1c617fcbc8c-Paper.pdf"

    else: continue

    paper_name = '[{}] '.format(paper_id) + paper.string + '.pdf'

    if ':' in paper_name:

    paper_name = paper_name.replace(':', '_')

    if '?' in paper_name:

    paper_name = paper_name.replace('?', '')

    paper_path = os.path.join(category_path, paper_name)

    download_pdf(paper_link, paper_path)

    print("Successfully downloaded {}!".format(paper_name))

    paper_id += 1

    print("Successfully downloaded all the papers from catagory {}!".format(category_path))

    def _save_html(html_url, html_path):

    """Save requested HTML files

    Args:

    html_url (str): url of the HTML page to be saved

    html_path (str): save path of HTML file

    """

    html_file = rq.get(html_url, headers=headers)

    with open(html_path, "w", encoding='utf-8') as h:

    h.write(html_file.text)

    def download_key_papers(root_dir):

    """Download all the key papers, consistent with the categories listed on the website

    Args:

    root_dir (str): save path of all the downloaded papers

    """

    # 1. Get the html of Spinning Up

    spinningup_html = rq.get(spinningup_url, headers=headers)

    # 2. Parse the html and get the main category ids

    soup = bf(spinningup_html.content, 'lxml')

    # _save_html(spinningup_url, 'spinningup.html')

    # spinningup_file = open('spinningup.html', 'r', encoding="UTF-8")

    # spinningup_handle = spinningup_file.read()

    # soup = bf(spinningup_handle, features='lxml')

    category_ids = []

    categories = soup.find(name='div', attrs={'class': 'section', 'id': 'key-papers-in-deep-rl'}).\

    find_all(name='div', attrs={'class': 'section'}, recursive=False)

    for category in categories:

    category_ids.append(category['id'])

    # 3. Get all the categories and make corresponding dirs

    category_dirs = []

    if not os.path.exitis(root_dir):

    os.makedirs(root_dir)

    for category in soup.find_all(name='h4'):

    category_name = list(category.children)[0].string

    if ':' in category_name: # replace ':' with '_' to get valid dir name

    category_name = category_name.replace(':', '_')

    category_path = os.path.join(root_dir, category_name)

    category_dirs.append(category_path)

    if not os.path.exists(category_path):

    os.makedirs(category_path)

    # 4. Start to download all the papers

    print("Start to download key papers...")

    for i in range(len(category_ids)):

    category_path = category_dirs[i]

    category_id = category_ids[i]

    content = soup.find(name='div', attrs={'class': 'section', 'id': category_id})

    inner_categories = content.find_all('div')

    if inner_categories != []:

    for category in inner_categories:

    category_id = category['id']

    inner_category = category.h4.text[:-1]

    inner_category_path = os.path.join(category_path, inner_category)

    if not os.path.exists(inner_category_path):

    os.makedirs(inner_category_path)

    content = soup.find(name='div', attrs={'class': 'section', 'id': category_id})

    papers = content.find_all(name='a',attrs={'class': 'reference external'})

    download_from_bs4(papers, inner_category_path)

    else:

    papers = content.find_all(name='a',attrs={'class': 'reference external'})

    download_from_bs4(papers, category_path)

    print("Download Complete!")

    if __name__ == "__main__":

    root_dir = "key-papers"

    download_key_papers(root_dir)

    1

    2

    3

    4

    5

    6

    7

    8

    9

    10

    11

    12

    13

    14

    15

    16

    17

    18

    19

    20

    21

    22

    23

    24

    25

    26

    27

    28

    29

    30

    31

    32

    33

    34

    35

    36

    37

    38

    39

    40

    41

    42

    43

    44

    45

    46

    47

    48

    49

    50

    51

    52

    53

    54

    55

    56

    57

    58

    59

    60

    61

    62

    63

    64

    65

    66

    67

    68

    69

    70

    71

    72

    73

    74

    75

    76

    77

    78

    79

    80

    81

    82

    83

    84

    85

    86

    87

    88

    89

    90

    91

    92

    93

    94

    95

    96

    97

    98

    99

    100

    101

    102

    103

    104

    105

    106

    107

    108

    109

    110

    111

    112

    113

    114

    115

    116

    117

    118

    119

    120

    121

    122

    123

    124

    125

    126

    127

    128

    129

    130

    131

    132

    133

    134

    135

    136

    137

    138

    139

    140

    141

    142

    143

    144

    145

    146

    147

    1

    2

    3

    4

    5

    6

    7

    8

    9

    10

    11

    12

    13

    14

    15

    16

    17

    18

    19

    20

    21

    22

    23

    24

    25

    26

    27

    28

    29

    30

    31

    32

    33

    34

    35

    36

    37

    38

    39

    40

    41

    42

    43

    44

    45

    46

    47

    48

    49

    50

    51

    52

    53

    54

    55

    56

    57

    58

    59

    60

    61

    62

    63

    64

    65

    66

    67

    68

    69

    70

    71

    72

    73

    74

    75

    76

    77

    78

    79

    80

    81

    82

    83

    84

    85

    86

    87

    88

    89

    90

    91

    92

    93

    94

    95

    96

    97

    98

    99

    100

    101

    102

    103

    104

    105

    106

    107

    108

    109

    110

    111

    112

    113

    114

    115

    116

    117

    118

    119

    120

    121

    122

    123

    124

    125

    126

    127

    128

    129

    130

    131

    132

    133

    134

    135

    136

    137

    138

    139

    140

    141

    142

    143

    144

    145

    146

    147

    二、openai写的文章能直接用吗

    openai写的文章能直接用吗?

    回答是:一般情况下上面写的文章是不可以直接用的因为上面写的文章也是从网络上一些文章上面拼凑而来的所以说你要自己的进行一个修改才能够使用

    三、chatgpt怎么下载?

    ChatGPT 是一个基于云的自然语言处理服务,无需下载。您可以在支持 ChatGPT 的应用程序或网站上直接使用它。一些应用程序或网站可能需要您创建账户或登录,但您不需要下载 ChatGPT 或其他任何软件或工具。

    如果您想在自己的应用程序或网站中集成 ChatGPT 的功能,您可以访问 OpenAI 的官方网站,注册 OpenAI API 的开发者账户,并按照指南获取 API 密钥。请注意,OpenAI API 目前处于测试阶段,需要提交应用程序才能获得 API 密钥。有关如何使用 OpenAI API 和集成 ChatGPT 的更多信息,请参阅 OpenAI 的文档和开发者指南。

    对于无法使用的用户可以查询whysou网看看,com上面有方法可以访问.

    四、openai本地无法使用

    openai本地无法使用万是因为它这个系统不兼容,所以需要把系统进行更新之后才可以使用,因为这是一款隐藏文的,所以如果这个文档被隐藏的话才会无法使用。

    以上就是关于openai免费使用相关问题的回答。希望能帮到你,如有更多相关问题,您也可以联系我们的客服进行咨询,客服也会为您讲解更多精彩的知识和内容。


    推荐阅读:

    在线修图编辑器免费(photoshop在线修图)

    手机按键在哪里设置vivo(手机按键在哪里设置OPPO)

    怎样自定义桌面图标手机oppo(oppo如何自定义桌面图标)

    设计商标图案(logo设计)

    贵州日式景观设计施工公司(贵州日式景观设计施工公司排名)