prompt
stringlengths
19
1.03M
completion
stringlengths
4
2.12k
api
stringlengths
8
90
#!/usr/bin/python3 import sys import copy from pathlib import Path from datetime import datetime,timedelta import re import matplotlib.pyplot as plt import math import numpy as np import random import pandas as pd import subprocess from pickle import dump,load from predictor.utility import msg2log from clustgelDL.au...
pd.Timestamp.now()
pandas.Timestamp.now
#!/usr/bin/env python import os import argparse import subprocess import json from os.path import isfile, join, basename import time import pandas as pd from datetime import datetime import tempfile import sys sys.path.append( os.path.abspath(os.path.join(os.path.dirname(__file__), os.path.pardir, 'instance_gene...
pd.DataFrame(results)
pandas.DataFrame
import os from typing import List, Tuple, Union import numpy as np import pandas as pd DATASET_DIR: str = "data/" # https://www.kaggle.com/rakannimer/air-passengers def read_air_passengers() -> Tuple[pd.DataFrame, np.ndarray]: indexes = [6, 33, 36, 51, 60, 100, 135] values = [205, 600, 150, 315, 150, 190, 6...
pd.read_csv(f"{DATASET_DIR}air_passengers.csv")
pandas.read_csv
#!/usr/bin/env python # -*- encoding: utf-8 -*- ''' @File : ioutil.py @Desc : Input and output data function. ''' # here put the import lib import os import sys import pandas as pd import numpy as np from . import TensorData import csv from .basicutil import set_trace class File(): def __init__(self,...
pd.DataFrame()
pandas.DataFrame
import logging import os import pickle import tarfile from typing import Tuple import numpy as np import pandas as pd import scipy.io as sp_io import shutil from scipy.sparse import csr_matrix, issparse from scMVP.dataset.dataset import CellMeasurement, GeneExpressionDataset, _download logger = logging.getLogger(__n...
pd.DataFrame(self.ATAC_name)
pandas.DataFrame
from flask import Flask, render_template, jsonify, request from flask_pymongo import PyMongo from flask_cors import CORS, cross_origin import json import copy import warnings import re import pandas as pd pd.set_option('use_inf_as_na', True) import numpy as np from joblib import Memory from xgboost import XGBClassi...
pd.concat([DataRows2, hotEncoderDF2], axis=1)
pandas.concat
import pandas as pd import numpy as np import json PROCESS_FILE_NAME_LIST = ["taxi_sort_01", "taxi_sort_001", "taxi_sort_002", "taxi_sort_003", "taxi_sort_004", "taxi_sort_005", "taxi_sort_006", "taxi_sort_007", "taxi_sort_008", "taxi_sort_009", "taxi_sort_0006", "taxi_sort_0007", "taxi_sort_0008", "taxi_sort_0009"] P...
pd.read_csv("precinct_center.csv", index_col=False)
pandas.read_csv
import pandas as pd import numpy as np import matplotlib.pyplot as plt plt.style.use('ggplot') target = 'scale' # IP plot_mode = 'all_in_one' obj = 'occ' # Port flow_dir = 'all' port_dir = 'sys' user_plot_pr = ['TCP'] user_plot_pr = ['UDP'] port_hist = pd.DataFrame({'A' : []}) user_port_hist = pd.DataFrame({'A' : []...
pd.read_csv("./postprocessed_data/%s/%s_piece%d.csv" % (files[data_idx], files[data_idx], piece_idx), index_col=None, header=0)
pandas.read_csv
# %% [markdown] # This python script takes audio files from "filedata" from sonicboom, runs each audio file through # Fast Fourier Transform, plots the FFT image, splits the FFT'd images into train, test & validation # and paste them in their respective folders # Import Dependencies import numpy as np import pandas...
pd.DataFrame()
pandas.DataFrame
''' The analysis module Handles the analyses of the info and data space for experiment evaluation and design. ''' from slm_lab.agent import AGENT_DATA_NAMES from slm_lab.env import ENV_DATA_NAMES from slm_lab.lib import logger, util, viz import numpy as np import os import pandas as pd import pydash as ps import shutil...
pd.concat(session_fitness_data, axis=1)
pandas.concat
#!/usr/bin/env python3 # Project : From geodynamic to Seismic observations in the Earth's inner core # Author : <NAME> """ Implement classes for tracers, to create points along the trajectories of given points. """ import numpy as np import pandas as pd import math import matplotlib.pyplot as plt from . import data...
pd.DataFrame(data=self.velocity_gradient, columns=["dvx/dx", "dvx/dy", "dvx/dz", "dvy/dx", "dvy/dy", "dvy/dz", "dvz/dx", "dvz/dy", "dvz/dz"])
pandas.DataFrame
#!/usr/bin/env python import sys, time, code import numpy as np import pickle as pickle from pandas import DataFrame, read_pickle, get_dummies, cut import statsmodels.formula.api as sm from sklearn.externals import joblib from sklearn.linear_model import LinearRegression from djeval import * def shell():...
get_dummies(yy_df[categorical_features])
pandas.get_dummies
import os import numpy as np import pandas as pd from numpy import abs from numpy import log from numpy import sign from scipy.stats import rankdata import scipy as sp import statsmodels.api as sm from data_source import local_source from tqdm import tqdm as pb # region Auxiliary functions def ts_sum(df, window=10): ...
pd.Series(result_industryaveraged_df.index)
pandas.Series
from turtle import TPen, color import numpy as np import pandas as pd import random import matplotlib.pyplot as plt import seaborn as sns import sklearn.metrics as metrics from keras.models import Sequential from keras.layers import Dense, LSTM, Flatten, Dropout def get_ace_values(temp_list): ''' This functio...
pd.DataFrame()
pandas.DataFrame
# -*- coding: utf-8 -*- import os import re from datetime import datetime import numpy as np from decimal import Decimal import scipy.io as sio import pandas as pd from tqdm import tqdm import glob from decimal import Decimal import datajoint as dj from pipeline import (reference, subject, acquisition, stimulation, ...
pd.concat([fixed_delay_xlsx, random_long_delay_xlsx, random_short_delay_xlsx, tactile_xlsx, sound12_xlsx])
pandas.concat
import sys import numpy as np import pandas as pd from loguru import logger from sklearn import model_selection from utils import dataset_utils default_settings = { 'data_definition_file_path': 'dataset.csv', 'folds_num': 5, 'data_random_seed': 1509, 'train_val_fraction': 0.8, 'trai...
pd.concat(groups_test_df_list)
pandas.concat
# -*- encoding: utf-8 -*- import functools import warnings from typing import List, Optional, Tuple, Union import numpy as np import pandas as pd from pandas.api.types import is_numeric_dtype import scipy.sparse import sklearn.utils from sklearn import preprocessing from sklearn.compose import make_column_transfor...
pd.api.types.is_numeric_dtype(x)
pandas.api.types.is_numeric_dtype
import functools import json import os from multiprocessing.pool import Pool from typing import List, Tuple, Type, Any import pprint import abc import luigi import numpy as np import pandas as pd import torch import torchbearer from torchbearer import Trial from tqdm import tqdm import gc from mars_gym.data.dataset im...
pd.read_csv(self.model_training.metadata_data_frame_path, dtype = {self.model_training.project_config.item_column.name : "str"})
pandas.read_csv
# -*- coding: utf-8 -*- """ Created on Tue Jul 27 10:23:59 2021 @author: alber """ import re import os import pandas as pd import numpy as np import spacy import pickle import lightgbm as lgb import imblearn from sklearn import preprocessing from sklearn.semi_supervised import ( LabelPropagation, LabelSpread...
pd.read_csv("train_dataset.csv")
pandas.read_csv
# import sys # sys.path.append('JEMIPYC') # from array_check_function_global import df,dfn,dfv,dfx,dfnx,dfvx import pandas as pd import numpy as np tab = '__' # no-extension , number of parameters is not limited, 2 or 3, whatever you want. # ex) df(A,B,C,D,...,Z...) # of course you just put one parameter....
pd.set_option('display.max_columns', None)
pandas.set_option
import os import pandas as pd import matplotlib.pyplot as plt import datapackage as dp import plotly.io as pio import plotly.offline as offline from plots import ( hourly_plot, stacked_plot, price_line_plot, price_scatter_plot, merit_order_plot, filling_level_plot, ) results = [r for r in os.l...
pd.concat([storages[r], shadow_prices[r]], axis=1)
pandas.concat
from datetime import datetime import numpy as np import pytest import pandas.util._test_decorators as td from pandas.core.dtypes.base import _registry as ea_registry from pandas.core.dtypes.common import ( is_categorical_dtype, is_interval_dtype, is_object_dtype, ) from pandas.core.dtypes.dtypes import (...
Series(sp_array, name="new_column")
pandas.Series
import numpy as np import pandas as pd from numba import njit from datetime import datetime import pytest from itertools import product from sklearn.model_selection import TimeSeriesSplit import vectorbt as vbt from vectorbt.generic import nb seed = 42 day_dt = np.timedelta64(86400000000000) df = pd.DataFrame({ ...
pd.DatetimeIndex(['2018-01-04'], dtype='datetime64[ns]', name='split_0', freq=None)
pandas.DatetimeIndex
import numpy as np import pandas as pd import spacy from spacy.lang.de.stop_words import STOP_WORDS from nltk.tokenize import sent_tokenize from itertools import groupby import copy import re import sys import textstat # Method to create a matrix with contains only zeroes and a index starting by 0 def cr...
pd.DataFrame(d_multi_word_list)
pandas.DataFrame
from __future__ import division import configparser import logging import os import re import time from collections import OrderedDict import numpy as np import pandas as pd import scipy.interpolate as itp from joblib import Parallel from joblib import delayed from matplotlib import pyplot as plt from pyplanscoring....
pd.DataFrame(res)
pandas.DataFrame
# -*- coding: utf-8 -*- # Author: <NAME> <<EMAIL>> # License: BSD """ Toolset working with yahoo finance data Module includes functions for easy access to YahooFinance data """ import urllib.request import numpy as np import requests # interaction with the web import os # file system operati...
pd.DataFrame(data,index=idx)
pandas.DataFrame
from __future__ import division from functools import wraps import pandas as pd import numpy as np import time import csv, sys import os.path import logging from .ted_functions import TedFunctions from .ted_aggregate_methods import TedAggregateMethods from base.uber_model import UberModel, ModelSharedInputs class Te...
pd.Series([], dtype="float", name="arbt_inv_sensory")
pandas.Series
import pandas as pd import json import streamlit as st import plotly.express as px from PIL import Image #Membuka file json with open ("kode_negara_lengkap.json") as file : data = json.load (file) #Menjadikan file json sebagai dataframe df_json = pd.DataFrame(data) #Membuka file csv df =
pd.read_csv('produksi_minyak_mentah.csv')
pandas.read_csv
import numpy as np import pandas as pd from sklearn.model_selection import train_test_split, StratifiedKFold from sklearn import svm from sklearn import metrics def classification(latent_code, random_seed=42, ten_fold=False): tumour_type =
pd.read_csv('data/PANCAN/GDC-PANCAN_both_samples_tumour_type.tsv', sep='\t', index_col=0)
pandas.read_csv
from datetime import datetime from io import StringIO import itertools import numpy as np import pytest import pandas.util._test_decorators as td import pandas as pd from pandas import ( DataFrame, Index, MultiIndex, Period, Series, Timedelta, date_range, ) import pandas._testing as tm ...
MultiIndex.from_product([["A"], [0, 1]])
pandas.MultiIndex.from_product
from flowsa.common import WITHDRAWN_KEYWORD from flowsa.flowbyfunctions import assign_fips_location_system from flowsa.location import US_FIPS import math import pandas as pd import io from flowsa.settings import log from string import digits YEARS_COVERED = { "asbestos": "2014-2018", "barite": "2014-2018", ...
pd.DataFrame()
pandas.DataFrame
import numpy as np import pandas as pd import pytest from rulelist.datastructure.attribute.nominal_attribute import activation_nominal, NominalAttribute class TestNominalAttribute(object): def test_normal(self): dictdata = {"column1" : np.array(["below50" if i < 50 else "above49" for i in range(100)]), ...
pd.testing.assert_series_equal(actual_vector, expected_vector, check_exact=True)
pandas.testing.assert_series_equal
#! -*- coding: utf-8 -*- from PIL import Image import matplotlib.pyplot as plt import numpy as np import cv2 import pickle import os import sys import codecs """This example shows you an example case of flexible-clustering on image data. In this example, it uses sub data from cifar-10 image collection. The clustering ...
pandas.DataFrame(table_objects['cluster_information'])
pandas.DataFrame
from tensorflow.keras.callbacks import TensorBoard import tensorflow as tf import keras from keras.datasets import mnist from keras.models import Sequential from keras.layers import Dense, Dropout, Activation from keras import backend as K from sqlalchemy import create_engine from sklearn.preprocessing import StandardS...
pd.read_sql_table(curr_Pair, engine)
pandas.read_sql_table
# %% import os import sys from collections import Counter from datetime import datetime, timedelta from glob import glob from pathlib import Path from zipfile import ZipFile # data wrangling import geopandas as gpd import pandas as pd import numpy as np import requests from urllib.error import HTTPError # data maniupl...
pd.set_option('display.max_rows', 5)
pandas.set_option
# coding:utf-8 # # The MIT License (MIT) # # Copyright (c) 2016-2020 # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, c...
pd.to_datetime(_data['date'])
pandas.to_datetime
# -*- coding: utf-8 -*- """ @author: HYPJUDY 2019/4/15 https://github.com/HYPJUDY Decoupling Localization and Classification in Single Shot Temporal Action Detection ----------------------------------------------------------------------------------- Operations used by Decouple-SSAD """ import pandas as pd import ...
pd.concat([resultDf1, resultDf2])
pandas.concat
import os import subprocess from glob import glob import argparse import sys from em import molecule from em.dataset import metrics from mpi4py import MPI from mpi4py.futures import MPICommExecutor from concurrent.futures import wait from scipy.spatial import cKDTree import numpy as np import pandas as pd import trace...
pd.DataFrame(columns=['id','map_path','contourLevel','subunit', 'tagged_path', 'number_points','tagged_points_path'])
pandas.DataFrame
""" Evaluates the model. """ import argparse import matplotlib as mpl # do not use Qt/X that require $DISPLAY, must be called before importing pyplot mpl.use('Agg') import numpy as np import pandas as pd from sklearn.metrics import confusion_matrix from prepare_training_data import load_indexes, load_transformers imp...
pd.read_csv(evaluation_dir + '/final_metrics.csv', index_col=0)
pandas.read_csv
"""Тесты для таблицы с торгуемыми ценными бумагами.""" from datetime import date import pandas as pd import pytest from poptimizer.data import ports from poptimizer.data.domain import events from poptimizer.data.domain.tables import base, securities from poptimizer.shared import col TICKER_CASES = ( ("GAZP", 0),...
pd.DataFrame([1, 4], index=["AKRN", "RTKMP"])
pandas.DataFrame
# Copyright (c) 2019, MD2K Center of Excellence # - <NAME> <<EMAIL>>, <NAME> <<EMAIL>> # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above co...
pd.DataFrame([], columns=column_names)
pandas.DataFrame
#!/usr/bin/python3 # -*- coding: utf-8 -*- import arrow import pandas as pd import requests import json from functools import reduce # RU-1: European and Uralian Market Zone (Price Zone 1) # RU-2: Siberian Market Zone (Price Zone 2) # RU-AS: Russia East Power System (2nd synchronous zone) # Handling of hours: data at...
pd.DataFrame(data)
pandas.DataFrame
import pandas as pd from primus.category import OneHotEncoder def test_fit_transform_HaveHandleUnknownValueAndUnseenValues_ExpectAllZeroes(): train = pd.DataFrame({'city': ['Chicago', 'Seattle']}) test = pd.DataFrame({'city': ['Chicago', 'Detroit']}) expected_result = pd.DataFrame({'city_1': [1, 0], ...
pd.testing.assert_frame_equal(expected_result, result)
pandas.testing.assert_frame_equal
from selenium import webdriver from selenium.webdriver.chrome.options import Options from selenium.webdriver.common.keys import Keys import requests import time from datetime import datetime import pandas as pd from urllib import parse from config import ENV_VARIABLE from os.path import getsize fold_path = ...
pd.DataFrame()
pandas.DataFrame
""" dataset = AbstractDataset() """ from collections import OrderedDict, defaultdict import json from pathlib import Path import numpy as np import pandas as pd from tqdm import tqdm import random def make_perfect_forecast(prices, horizon): prices = np.array(prices).reshape(-1, 1) forecast = np.hstack([n...
pd.concat(ds['features'], axis=1)
pandas.concat
import os import tempfile from StringIO import StringIO import numpy as np import numpy.testing as npt import pandas as pd import pytest import statsmodels.formula.api as smf import yaml from pandas.util import testing as pdt from statsmodels.regression.linear_model import RegressionResultsWrapper from .. import reg...
pdt.assert_series_equal(wrapper.bse, fit.bse, check_names=False)
pandas.util.testing.assert_series_equal
import matplotlib.pyplot as plt import os import seaborn as sns import numpy as np from matplotlib.colors import ListedColormap import pandas as pd from sklearn.manifold import TSNE from src.Utils.Fitness import Fitness class Graphs: def __init__(self,objectiveNames,data,save=True,display=False,path='./Figures/'...
pd.DataFrame(data, columns=['algorithm', 'nbRules','support','confidence','cosine'])
pandas.DataFrame
#!/usr/bin/env python # Copyright 2020 ARC Centre of Excellence for Climate Extremes # author: <NAME> <<EMAIL>> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/l...
pd.DataFrame(d)
pandas.DataFrame
import re from collections import defaultdict import pandas as pd def empty_data(mapping_rows: list, row: pd.DataFrame): return all([True if mapping_row not in row or row.get(mapping_row)=='' else False for mapping_row in mapping_rows]) def merge_spreadsheets(workbook: str, merge_field: str) -> pd.DataFrame...
pd.read_excel(workbook, sheet_name=None)
pandas.read_excel
# -*- coding:utf-8 -*- # /usr/bin/env python """ Date: 2020/3/17 13:06 Desc: 期货-中国-交易所-会员持仓数据接口 大连商品交易所、上海期货交易所、郑州商品交易所、中国金融期货交易所 采集前 20 会员持仓数据; 建议下午 16:30 以后采集当天数据, 避免交易所数据更新不稳定; 郑州商品交易所格式分为三类 大连商品交易所有具体合约的持仓排名, 通过 futures_dce_position_rank 获取 20171228 http://www.czce.com.cn/cn/DFSStaticFiles/Future/2020/20200727/Futu...
pd.read_html(r.text)
pandas.read_html
# %% from selenium import webdriver from selenium.webdriver.common.by import By from selenium.webdriver.support.ui import WebDriverWait from selenium.webdriver.support import expected_conditions as EC import pandas as pd import sys from allensdk.api.queries.image_download_api import ImageDownloadApi from allensdk.con...
pd.read_csv(infile, index_col=0)
pandas.read_csv
#%% import numpy as np import pandas as pd from orderedset import OrderedSet as oset #%% wals = pd.read_csv('ISO_completos.csv').rename(columns={'Status':'Status_X_L'}) wals_2 = pd.read_csv('ISO_completos_features.csv').rename(columns={'Status':'Status_X_L'}) wiki_merged = pd.read_csv('Wikidata_Wals_IDWALS.csv') wiki =...
pd.concat(collapsed, axis=1)
pandas.concat
"""Project Palette functions for palette project """ from tkinter import filedialog from tkinter.constants import END import tkinter.messagebox as msgbox import os import webbrowser from bs4 import BeautifulSoup import pandas as pd from PIL import Image import requests if __name__ == "__main__": from data import...
pd.read_csv(filename, nrows=1)
pandas.read_csv
import os import sys import argparse import numpy as np import pandas as pd import cv2 import matplotlib.pyplot as plt from tqdm import tqdm import torch import torch.nn.functional as TF import torch.backends.cudnn as cudnn from torch.utils.data import DataLoader sys.path.append('../') # from torchlib.transforms i...
pd.DataFrame(tuplas)
pandas.DataFrame
import json import pandas as pd import argparse #Test how many points the new_cut_dataset has parser = argparse.ArgumentParser() parser.add_argument('--dataset_path', default="new_dataset.txt", type=str, help="Full path to the txt file containing the dataset") parser.add_argument('--discretization_unit', default=1, typ...
pd.to_datetime(data['start_date'])
pandas.to_datetime
import os import sys import joblib # sys.path.append('../') main_path = os.path.split(os.getcwd())[0] + '/covid19_forecast_ml' import numpy as np import pandas as pd import matplotlib.pyplot as plt import seaborn as sns from datetime import datetime, timedelta from tqdm import tqdm from Dataloader_v2 import BaseCOVDa...
pd.to_datetime(data_cases['date_time'], format='%Y-%m-%d')
pandas.to_datetime
# -*- coding: utf-8 -*- """ This module is designed for the use with the coastdat2 weather data set of the Helmholtz-Zentrum Geesthacht. A description of the coastdat2 data set can be found here: https://www.earth-syst-sci-data.net/6/147/2014/ SPDX-FileCopyrightText: 2016-2019 <NAME> <<EMAIL>> SPDX-License-Identifi...
pd.DataFrame()
pandas.DataFrame
import pandas as pd import os def _1996(data_dir): from . import sgf_table_sums file = "96data35.txt" ids = pd.read_excel( os.path.join(data_dir, "government-ids.xls"), dtype={"ID Code": str, "State": str}, ) ids["State"] = ids["State"].str.strip() map_id = dict(zip(ids["ID...
pd.DataFrame(columns=cols)
pandas.DataFrame
import pandas as pd import numpy as np import matplotlib.pyplot as plt import seaborn as sns import os.path import math from IPython.display import display,clear_output import random import scipy.stats as st from sklearn.preprocessing import LabelEncoder import sklearn.preprocessing as sk import sklearn....
pd.DataFrame()
pandas.DataFrame
import argparse import numpy as np import pandas import utils parser = argparse.ArgumentParser() parser.add_argument("data_path", type=str, help="path to csv file") utils.add_arguments(parser, ["output"]) args = parser.parse_args() data_path = args.data_path out_path = args.output df = pandas.read_csv(data_path) aggr...
pandas.DataFrame(aggregate_dict)
pandas.DataFrame
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Description ---------- Some simple classes to be used in sklearn pipelines for pandas input Informations ---------- Author: <NAME> Maintainer: Email: <EMAIL> Copyright: Credits: License: Version: Status: in development """ import...
pd.concat(list_df, 1)
pandas.concat
import pandas as pd import numpy as np from sklearn.datasets import load_breast_cancer as lbc from tkinter import * from tkinter import messagebox data = lbc() clm = np.array(data['feature_names']) df_x = pd.DataFrame(data['data']) df_y =
pd.DataFrame(data['target'])
pandas.DataFrame
from datetime import datetime import numpy as np import pandas as pd import pytest from numba import njit import vectorbt as vbt from tests.utils import record_arrays_close from vectorbt.generic.enums import range_dt, drawdown_dt from vectorbt.portfolio.enums import order_dt, trade_dt, log_dt day_dt = np.timedelta64...
pd.testing.assert_index_equal(stats_df.index, ranges.wrapper.columns)
pandas.testing.assert_index_equal
from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import sys import copy from datetime import datetime import time import pickle import random import pandas as pd import numpy as np import tensorflow as tf import pathlib from sklearn import preproces...
pd.to_datetime(self.config.end_date, format="%Y%m%d")
pandas.to_datetime
import pandas as pd import numpy as np import math import matplotlib.pyplot as plt import assetallocation_arp.models.ARP as arp # Parameters TIMES_LAG=3 settings=arp.dataimport_settings("Settings") # Change the universe of markets that is being used markets="Leverage_MATR" # All "Leverage_all_markets" / Minimalist ...
pd.read_pickle("Data.pkl")
pandas.read_pickle
from __future__ import print_function import unittest from unittest import mock from io import BytesIO, StringIO import random import six import os import re import logging import numpy as np import pandas as pd from . import utils as test_utils import dataprofiler as dp from dataprofiler.profilers.profile_builder ...
pd.read_csv(cls.input_file_path)
pandas.read_csv
# -*- coding: utf-8 -*- import pytest import numpy as np import pandas as pd import pandas.util.testing as tm import pandas.compat as compat ############################################################### # Index / Series common tests which may trigger dtype coercions ###############################################...
pd.Series([1, 2, 3, 4])
pandas.Series
import pandas as pd import os,sys import re import torch inp_path = r'/home/tiwarikajal/embeddingdata' out_path = r'/home/tiwarikajal/data/' error = [] df =
pd.DataFrame(columns=['year', 'Company', 'embeddings1a', 'embeddings7'])
pandas.DataFrame
import mysql.connector import pandas as pd class MySQLInterface: def __init__(self, server, username, password, dbname): self.server = server self.username = username self.password = password self.dbname = dbname def __connect(self): try: ...
pd.DataFrame(output)
pandas.DataFrame
import pandas as pd def generate_train(playlists): # define category range cates = {'cat1': (10, 50), 'cat2': (10, 78), 'cat3': (10, 100), 'cat4': (40, 100), 'cat5': (40, 100), 'cat6': (40, 100),'cat7': (101, 250), 'cat8': (101, 250), 'cat9': (150, 250), 'cat10': (150, 250)} cat_pids = {} ...
pd.concat([df_test_itr, df_sample])
pandas.concat
#!/usr/bin/env python3 import os import io import re import argparse import itertools import collections as col import operator as op import pickle as pck import pandas as pd def parse_args(): parser = argparse.ArgumentParser() parser.add_argument( '--agp-file', '-a', type=str, ...
pd.read_csv(bed_path, sep='\t', names=bed_columns, header=None)
pandas.read_csv
#!/usr/bin/env python # coding: utf-8 # # **<NAME> - Tracking Data Assignment** # # Sunday 11th October 2020 # # --- # In[1]: import pandas as pd import numpy as np import datetime # imports required by data prep functions import json # Laurie's libraries import scipy.signal as signal import matplotlib.animation ...
pd.DataFrame(homePlayers)
pandas.DataFrame
# -*- coding: utf-8 -*- ''' TopQuant-TQ极宽智能量化回溯分析系统2019版 Top极宽量化(原zw量化),Python量化第一品牌 by Top极宽·量化开源团队 2019.01.011 首发 网站: www.TopQuant.vip www.ziwang.com QQ群: Top极宽量化总群,124134140 文件名:toolkit.py 默认缩写:import topquant2019 as tk 简介:Top极宽量化·常用量化系统参数模块 ''' # import sys, os, re import arrow, bs4, rando...
pd.to_datetime(df.index, format='%Y-%m-%dT%H:%M:%S')
pandas.to_datetime
import numpy as np import pandas as pd from tqdm import tqdm from prereise.gather.solardata.helpers import get_plant_id_unique_location from prereise.gather.solardata.nsrdb.nrel_api import NrelApi def retrieve_data(solar_plant, email, api_key, year="2016"): """Retrieve irradiance data from NSRDB and calculate th...
pd.DataFrame({"Pout": [], "plant_id": [], "ts": [], "ts_id": []})
pandas.DataFrame
#################### # Import Libraries #################### import os import sys from PIL import Image import cv2 import numpy as np import pandas as pd import pytorch_lightning as pl from pytorch_lightning.metrics import Accuracy from pytorch_lightning import loggers from pytorch_lightning import seed_e...
pd.DataFrame()
pandas.DataFrame
import gradio as gr import pickle import os import pandas as pd import json import urllib.parse from stats import create_pdf from pycaret.classification import * welcome_message = """ Hello ! Thanks for using our tool , you'll be able to build your own recommandation tool. You'll be able...
pd.concat([liked, bad1, bad2, bad3, bad4])
pandas.concat
import web import pandas as pd import numpy as np import common import os import click def hydro_op_chars_inputs_(webdb, project, hydro_op_chars_sid, balancing_type_project): rows = webdb.where("inputs_project_hydro_operational_chars", proj...
pd.read_excel(datapath, sheet_name=project, nrows=365)
pandas.read_excel
"""Interactions with rainfall and river data.""" import numpy as np import pandas as pd __all__ = ["get_station_data"] def get_station_data(filename, station_reference): """Return readings for a specified recording station from .csv file. Parameters ---------- filename: str filename to rea...
pd.read_csv(filename)
pandas.read_csv
import datetime import pandas as pd from pathlib import Path import matplotlib.pyplot as plt _repos_csv = [] _issues_csv = [] CSV_FPATH = Path('/home/lucas.rotsen/Git_Repos/benchmark_frameworks/github_metrics') METRICS_FPATH = Path('/home/lucas.rotsen/Git_Repos/benchmark_frameworks/metrics/raw') def load_csv(file):...
pd.concat(dfs)
pandas.concat
# Test for evaluering af hvert forecast og sammenligning mellem forecast import pandas as pd import numpy as np from numpy.random import rand from numpy import ix_ from itertools import product import chart_studio.plotly as py import chart_studio import plotly.graph_objs as go import statsmodels.api as sm chart_studio...
pd.DataFrame(eq)
pandas.DataFrame
from __future__ import annotations import logging import os import numpy as np import json import warnings import sys import shutil from datetime import timedelta import pandas as pd import pickle import copy import yaml import torch from torch import nn from torch.nn.modules.loss import _Loss import torch.nn.functiona...
pd.DataFrame(to_be_converted, index=index, columns=self.class_labels)
pandas.DataFrame
# Ref: https://towardsdatascience.com/data-apps-with-pythons-streamlit-b14aaca7d083 #/app.py import streamlit as st import json import requests # import sys # import os import pandas as pd import numpy as np import re from datetime import datetime as dt from pandas_profiling import ProfileReport from streamlit_pandas...
pd.DataFrame(msg_dict)
pandas.DataFrame
import pandas as pd import random import math import numpy as np import matplotlib.pyplot as plt from shapely.geometry.polygon import LinearRing, Polygon, Point from maxrect import get_intersection, get_maximal_rectangle, rect2poly from vertical_adhesion import * def get_min_max(input_list): ''' get minimum a...
pd.DataFrame(stitches_per_layer, columns=['layer', 'stitch'])
pandas.DataFrame
#!/usr/bin/env python # coding: utf-8 import numpy as np import pandas as pd from copy import deepcopy from functools import partial import matplotlib.pyplot as plt import optuna import pickle from sklearn.metrics import mean_squared_error from tqdm import tqdm import os code_path = os.path.dirname(os.path.abspath...
pd.concat(pred_new)
pandas.concat
import pytz import pytest import dateutil import warnings import numpy as np from datetime import timedelta from itertools import product import pandas as pd import pandas._libs.tslib as tslib import pandas.util.testing as tm from pandas.errors import PerformanceWarning from pandas.core.indexes.datetimes import cdate_...
DatetimeIndex(['2011-01-01'], freq='D')
pandas.DatetimeIndex
# -*- coding: utf-8 -*- """ __date: 2021/05/12 __author: ssfang __corporation: OriginQuantum __usage: """ import os import re import time import threading from queue import Queue from datetime import datetime import yaml import pandas as pd import matplotlib.pyplot as plt pd.set_option('display.max_c...
pd.set_option('display.max_rows', None)
pandas.set_option
import os import sys import pdb import bdb import click import logging import signal import hashlib import inspect import traceback import pandas as pd from subir import Uploader from .browser_interactor import BrowserInteractor from .user_interactor import UserInteractor, Interaction from .pilot import Pilot from .ma...
pd.DataFrame()
pandas.DataFrame
import numpy as np import pandas as pd import pytest import orca from urbansim_templates import utils def test_parse_version(): assert utils.parse_version('0.1.0.dev0') == (0, 1, 0, 0) assert utils.parse_version('0.115.3') == (0, 115, 3, None) assert utils.parse_version('3.1.dev7') == (3, 1, 0, 7) a...
pd.Series([10,5], index=[3,1])
pandas.Series
import PyPDF2 import csv from pathlib import Path import io import pandas import numpy from pdfminer.pdfinterp import PDFResourceManager, PDFPageInterpreter from pdfminer.converter import TextConverter from pdfminer.layout import LAParams from pdfminer.pdfpage import PDFPage def Cpk(usl, lsl, avg, sigma , cf,...
pandas.set_option('display.expand_frame_repr', False)
pandas.set_option
import pandas as pd from evaluate.calculator import ( RecallCalculator, PrecisionCalculator, EmptyReportError, ) import pytest from unittest.mock import patch, Mock from evaluate.report import ( Report, PrecisionReport, RecallReport ) from tests.common import create_precision_report_row from io ...
pd.DataFrame(columns=columns)
pandas.DataFrame
from set_figure_defaults import FigureDefaults import numpy as np import matplotlib.pyplot as plt import pandas as pd import seaborn as sn from sklearn import preprocessing from sklearn.model_selection import train_test_split from sklearn.ensemble import RandomForestRegressor import operator import warnings import pick...
pd.DataFrame(dataset.loc[:,corrMatrixFiltered.columns[0]])
pandas.DataFrame
import numpy as np import pandas as pd import pandas.util.testing as tm import pytest import pandas_datareader.data as web pytestmark = pytest.mark.stable class TestEurostat(object): def test_get_ert_h_eur_a(self): # Former euro area national currencies vs. euro/ECU # annual data (ert_h_eur_a) ...
pd.Timestamp("2000-01-01")
pandas.Timestamp
"""Module for running decoding experiments.""" from pathlib import Path from typing import Optional, Sequence, Union import numpy as np import pandas as pd from joblib import Parallel, delayed from sklearn.model_selection import BaseCrossValidator import pte_decode def run_experiment( feature_root: Union[Path, ...
pd.concat(features, axis=1)
pandas.concat
# -*- coding: utf-8 -*- import logging import os import click import pandas as pd from src.libs.bookmaker import BookMaker from sqlalchemy import create_engine import pymysql pymysql.install_as_MySQLdb() @click.command() @click.option('--model', default='mlp_1') @click.option('--strategy', default='value_bet_0.5') def...
pd.read_sql(sql="select MATCH_ID, BbAvH, BbAvD, BbAvA, FTR from matches", con=db)
pandas.read_sql
# Do some analytics on Shopify transactions. import pandas as pd from datetime import datetime, timedelta class Analytics: def __init__(self, filename: str, datetime_now, refund_window: int): raw = pd.read_csv(filename) clean = raw[raw['Status'].isin(['success'])] # Filter down t...
pd.merge(sales, total_refunds, on='Name', how='outer')
pandas.merge
#web scrapping libraries from bs4 import BeautifulSoup as bs import requests from selenium import webdriver from webdriver_manager.chrome import ChromeDriverManager from selenium.webdriver.chrome.options import Options #data processing libraries import fsspec import os import folium import time import numpy as np impor...
pd.DataFrame(data=data, columns=columns)
pandas.DataFrame
# -*- coding: utf-8 -*- """ Created on Thu Feb 18 14:22:56 2021 @author: KRS1BBH """ from ImportFilter import Importfile import pandas as pd import os, glob #get path of directory script is executed from dirname = os.path.dirname(__file__) #nuk Filelist=[dirname+'/testdata/NuK/LotResultSummaryAll.csv']...
pd.DataFrame()
pandas.DataFrame
# -*- coding: utf-8 -*- import logging import numpy from pandas import DataFrame, Series, pivot_table from scipy.spatial.distance import pdist, squareform from scipy.cluster.hierarchy import * from matplotlib import pyplot from pylie.methods.methods import hlinkage_to_treematrix from pylie.plotting import plot_matri...
pivot_table(self.data, values=column, index=['case'], columns=['poses'])
pandas.pivot_table
"""Tests various time series functions which are used extensively in tcapy """ __author__ = 'saeedamen' # <NAME> / <EMAIL> # # Copyright 2017 Cuemacro Ltd. - http//www.cuemacro.com / @cuemacro # # See the License for the specific language governing permissions and limitations under the License. # import pandas as p...
pd.date_range(start='01 Jan 2018', end='05 Jan 2018', freq='30s')
pandas.date_range
# -*- coding: utf-8 -*- """ Created on Thu Feb 16 23:11:56 2017 @author: Flamingo """ import pandas as pd import numpy as np import datetime import copy import sys sys.path.append('../TOOLS') from IJCAI2017_TOOL import * #%% readin shop data HOLI = pd.read_csv('../additional/HOLI.csv') HOLI = HOLI.set_index(['DAT...
pd.DataFrame(train_date_zip)
pandas.DataFrame