-
Notifications
You must be signed in to change notification settings - Fork 1
/
OllaBench1_gui.py
53 lines (45 loc) · 3.08 KB
/
OllaBench1_gui.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
import hmac
import streamlit as st
from OllaBench_gui_menu import menu, show_header
st.set_page_config(
page_title="OllaBench1 v.0.2",
initial_sidebar_state="expanded"
)
# Initialize st.session_state
if "role" not in st.session_state:
st.session_state.role = None
if "password" not in st.session_state:
st.session_state.password = None
if "username" not in st.session_state:
st.session_state.username = None
if "healthcheck_passed" not in st.session_state:
st.session_state.healthcheck_passed = False
# Main Streamlit app starts here
menu()
col1, col2 = st.columns([0.1,0.2])
with col1:
st.write(" ")
with col2:
st.image('logo.png', width=180)
st.markdown('''
<div align="center">
<h4> OllaBench is a novel a benchmark platform and dataset generator to evaluate Large Language Models' ability to reason and make decisions based on cognitive behavioral insights in cybersecurity. </h4>
[![Documentation](https://img.shields.io/badge/docs-latest-brightgreen.svg?style=flat)]()
[![python](https://img.shields.io/badge/python-3.10-green)]()
[![Static Badge](https://img.shields.io/badge/release-0.2-green?style=flat&color=green)]()
[![license](https://img.shields.io/badge/license-Apache%202-blue)](./LICENSE)
The grand challenge that most CEO's care about is maintaining the right level of cybersecurity at a minimum cost as companies are not able to reduce cybersecurity risks despite their increased cybersecurity investments [[1]](https://www.qbusiness.pl/uploads/Raporty/globalrisk2021.pdf). Fortunately, the problem can be explained via interdependent cybersecurity (IC) [[2]](https://www.nber.org/system/files/working_papers/w8871/w8871.pdf). Human factors account for half of the long-lasting challenges in IC as identified by Kianpour et al. [[3]](https://www.mdpi.com/2071-1050/13/24/13677), and Laszka et al. [[4]](http://real.mtak.hu/21924/1/Buttyan4.pdf). Unfortunately, human-centric research within the context of IC is under-explored while research on general IC has unrealistic assumptions about human factors. Fortunately, the dawn of Large Language Models (LLMs) promise a much efficient way to research and develop solutions to human-centric problems across domains. We also note that the Zero-trust principles require the evaluation, validation, and continuous monitoring of AI applications including LLMs.
Therefore, OllaBench was born to help both researchers and application developers conveniently evaluate their LLM models within the context of cybersecurity compliance or non-compliance behaviors. The main goal is to help with identifying the right LLM in the early phase of developing an LLM-based application.</div>
''',unsafe_allow_html=True)
st.markdown("**Introduction Video (90s)**")
col1, col2 = st.columns([0.04,0.9])
with col1:
st.write(" ")
with col2:
st.video('https://youtu.be/o-uIvInsOMA')
st.markdown("**Demonstration of use (3mins)**")
col1, col2 = st.columns([0.04,0.9])
with col1:
st.write(" ")
with col2:
st.video('https://youtu.be/aOEGLCnLo1M')