|
8 | 8 | import random |
9 | 9 | from scipy.stats import norm |
10 | 10 |
|
11 | | -# data = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/df2020.csv') |
12 | | -# df2018 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/df2018.csv') |
13 | | -# full_data2018 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/survey_results_sample_2018.csv') |
14 | | -# full_data2019=pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/survey_results_sample_2019.csv') |
15 | | -# full_df2020 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/survey_results_sample_2020.csv') |
16 | | -# df2019 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/df2019.csv') |
17 | | -# df2020 = data[(data['SalaryUSD'] < 200000)] |
| 11 | +data = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/df2020.csv') |
| 12 | +df2018 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/df2018.csv') |
| 13 | +full_data2018 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/survey_results_sample_2018.csv') |
| 14 | +full_data2019=pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/survey_results_sample_2019.csv') |
| 15 | +full_df2020 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/survey_results_sample_2020.csv') |
| 16 | +df2019 = pd.read_csv('https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/df2019.csv') |
| 17 | +df2020 = data[(data['SalaryUSD'] < 200000)] |
18 | 18 |
|
19 | | -# # features for job satisfaction |
20 | | -# results = pd.read_csv("https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/results.csv") |
| 19 | +# features for job satisfaction |
| 20 | +results = pd.read_csv("https://raw.githubusercontent.com/Recode-Hive/Stackoverflow-Analysis/main/streamlit/results.csv") |
21 | 21 |
|
22 | 22 |
|
23 | 23 | # Load the data from local CSV files |
24 | | -data = pd.read_csv('df2020.csv') |
25 | | -df2018 = pd.read_csv('df2018.csv') |
26 | | -full_data2018 = pd.read_csv('survey_results_sample_2018.csv') |
27 | | -full_data2019 = pd.read_csv('survey_results_sample_2019.csv') |
28 | | -full_df2020 = pd.read_csv('survey_results_sample_2020.csv') |
29 | | -df2019 = pd.read_csv('df2019.csv') |
30 | | - |
31 | | -# Filter the 2020 data |
32 | | -df2020 = data[(data['SalaryUSD'] < 200000)] |
| 24 | +# data = pd.read_csv('df2020.csv') |
| 25 | +# df2018 = pd.read_csv('df2018.csv') |
| 26 | +# full_data2018 = pd.read_csv('survey_results_sample_2018.csv') |
| 27 | +# full_data2019 = pd.read_csv('survey_results_sample_2019.csv') |
| 28 | +# full_df2020 = pd.read_csv('survey_results_sample_2020.csv') |
| 29 | +# df2019 = pd.read_csv('df2019.csv') |
| 30 | + |
| 31 | +# # Filter the 2020 data |
| 32 | +# df2020 = data[(data['SalaryUSD'] < 200000)] |
33 | 33 |
|
34 | | -# Load results for job satisfaction from the local file |
35 | | -results = pd.read_csv("results.csv") |
| 34 | +# # Load results for job satisfaction from the local file |
| 35 | +# results = pd.read_csv("results.csv") |
36 | 36 |
|
37 | 37 |
|
38 | 38 | ####################################### |
|
0 commit comments