Spaces:
Paused
Paused
| from requests import Session | |
| from uuid import uuid4 | |
| from json import loads | |
| import os | |
| import json | |
| import requests | |
| from ...typing import sha256, Dict, get_type_hints | |
| url = 'https://gpt-gm.h2o.ai' | |
| model = ['falcon-40b', 'falcon-7b', 'llama-13b'] | |
| supports_stream = True | |
| needs_auth = False | |
| models = { | |
| 'falcon-7b': 'h2oai/h2ogpt-gm-oasst1-en-2048-falcon-7b-v3', | |
| 'falcon-40b': 'h2oai/h2ogpt-gm-oasst1-en-2048-falcon-40b-v1', | |
| 'llama-13b': 'h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-13b' | |
| } | |
| def _create_completion(model: str, messages: list, stream: bool, **kwargs): | |
| conversation = 'instruction: this is a conversation beween, a user and an AI assistant, respond to the latest message, referring to the conversation if needed\n' | |
| for message in messages: | |
| conversation += '%s: %s\n' % (message['role'], message['content']) | |
| conversation += 'assistant:' | |
| client = Session() | |
| client.headers = { | |
| 'authority': 'gpt-gm.h2o.ai', | |
| 'origin': 'https://gpt-gm.h2o.ai', | |
| 'referer': 'https://gpt-gm.h2o.ai/', | |
| 'sec-ch-ua': '"Not.A/Brand";v="8", "Chromium";v="114", "Google Chrome";v="114"', | |
| 'sec-ch-ua-mobile': '?0', | |
| 'sec-ch-ua-platform': '"Windows"', | |
| 'sec-fetch-dest': 'document', | |
| 'sec-fetch-mode': 'navigate', | |
| 'sec-fetch-site': 'same-origin', | |
| 'sec-fetch-user': '?1', | |
| 'upgrade-insecure-requests': '1', | |
| 'user-agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/114.0.0.0 Safari/537.36', | |
| } | |
| client.get('https://gpt-gm.h2o.ai/') | |
| response = client.post('https://gpt-gm.h2o.ai/settings', data={ | |
| 'ethicsModalAccepted': 'true', | |
| 'shareConversationsWithModelAuthors': 'true', | |
| 'ethicsModalAcceptedAt': '', | |
| 'activeModel': 'h2oai/h2ogpt-gm-oasst1-en-2048-falcon-40b-v1', | |
| 'searchEnabled': 'true', | |
| }) | |
| headers = { | |
| 'authority': 'gpt-gm.h2o.ai', | |
| 'accept': '*/*', | |
| 'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3', | |
| 'origin': 'https://gpt-gm.h2o.ai', | |
| 'referer': 'https://gpt-gm.h2o.ai/', | |
| 'sec-ch-ua': '"Not.A/Brand";v="8", "Chromium";v="114", "Google Chrome";v="114"', | |
| 'sec-ch-ua-mobile': '?0', | |
| 'sec-ch-ua-platform': '"Windows"', | |
| 'sec-fetch-dest': 'empty', | |
| 'sec-fetch-mode': 'cors', | |
| 'sec-fetch-site': 'same-origin', | |
| 'user-agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/114.0.0.0 Safari/537.36', | |
| } | |
| json_data = { | |
| 'model': models[model] | |
| } | |
| response = client.post('https://gpt-gm.h2o.ai/conversation', | |
| headers=headers, json=json_data) | |
| conversationId = response.json()['conversationId'] | |
| completion = client.post(f'https://gpt-gm.h2o.ai/conversation/{conversationId}', stream=True, json = { | |
| 'inputs': conversation, | |
| 'parameters': { | |
| 'temperature': kwargs.get('temperature', 0.4), | |
| 'truncate': kwargs.get('truncate', 2048), | |
| 'max_new_tokens': kwargs.get('max_new_tokens', 1024), | |
| 'do_sample': kwargs.get('do_sample', True), | |
| 'repetition_penalty': kwargs.get('repetition_penalty', 1.2), | |
| 'return_full_text': kwargs.get('return_full_text', False) | |
| }, | |
| 'stream': True, | |
| 'options': { | |
| 'id': kwargs.get('id', str(uuid4())), | |
| 'response_id': kwargs.get('response_id', str(uuid4())), | |
| 'is_retry': False, | |
| 'use_cache': False, | |
| 'web_search_id': '' | |
| } | |
| }) | |
| for line in completion.iter_lines(): | |
| if b'data' in line: | |
| line = loads(line.decode('utf-8').replace('data:', '')) | |
| token = line['token']['text'] | |
| if token == '<|endoftext|>': | |
| break | |
| else: | |
| yield (token) | |
| params = f'g4f.Providers.{os.path.basename(__file__)[:-3]} supports: ' + \ | |
| '(%s)' % ', '.join([f"{name}: {get_type_hints(_create_completion)[name].__name__}" for name in _create_completion.__code__.co_varnames[:_create_completion.__code__.co_argcount]]) |