fix
Browse files
app.py
CHANGED
@@ -7,10 +7,14 @@ from huggingface_hub import InferenceClient, login
|
|
7 |
|
8 |
MAX_SEED = np.iinfo(np.int32).max
|
9 |
MAX_IMAGE_SIZE = 2048
|
|
|
10 |
|
|
|
|
|
|
|
11 |
|
12 |
-
def infer(
|
13 |
-
client = InferenceClient(provider="fal-ai", token=
|
14 |
image = client.text_to_image(
|
15 |
prompt=prompt,
|
16 |
width=width,
|
@@ -38,7 +42,8 @@ with gr.Blocks(css=css) as demo:
|
|
38 |
with gr.Sidebar():
|
39 |
gr.Markdown("# Inference Provider")
|
40 |
gr.Markdown("This Space showcases the black-forest-labs/FLUX.1-dev model, served by the nebius API. Sign in with your Hugging Face account to use this API.")
|
41 |
-
button = gr.LoginButton("Sign in")
|
|
|
42 |
with gr.Column(elem_id="col-container"):
|
43 |
gr.Markdown(f"""# FLUX.1 [schnell] with fal-ai through HF Inference Providers ⚡
|
44 |
learn more about HF Inference Providers [here](https://huggingface.co/docs/inference-providers/index)""")
|
|
|
7 |
|
8 |
MAX_SEED = np.iinfo(np.int32).max
|
9 |
MAX_IMAGE_SIZE = 2048
|
10 |
+
TOKEN = None
|
11 |
|
12 |
+
def get_token(oauth_token: gr.OAuthToken | None):
|
13 |
+
global TOKEN
|
14 |
+
TOKEN = oauth_token.token if oauth_token else None
|
15 |
|
16 |
+
def infer(prompt, seed=42, randomize_seed=False, width=1024, height=1024, num_inference_steps=4, progress=gr.Progress(track_tqdm=True)):
|
17 |
+
client = InferenceClient(provider="fal-ai", token=TOKEN)
|
18 |
image = client.text_to_image(
|
19 |
prompt=prompt,
|
20 |
width=width,
|
|
|
42 |
with gr.Sidebar():
|
43 |
gr.Markdown("# Inference Provider")
|
44 |
gr.Markdown("This Space showcases the black-forest-labs/FLUX.1-dev model, served by the nebius API. Sign in with your Hugging Face account to use this API.")
|
45 |
+
button = gr.LoginButton("Sign in", fn=get_token)
|
46 |
+
|
47 |
with gr.Column(elem_id="col-container"):
|
48 |
gr.Markdown(f"""# FLUX.1 [schnell] with fal-ai through HF Inference Providers ⚡
|
49 |
learn more about HF Inference Providers [here](https://huggingface.co/docs/inference-providers/index)""")
|