Update random bits and bobs
Signed-off-by: Nikolaos Karaolidis <nick@karaolidis.com>
This commit is contained in:
106
src/config.rs
106
src/config.rs
@@ -15,27 +15,6 @@ use rust_bert::{
|
||||
use std::{env, num::NonZeroU32, path::PathBuf, sync::Arc};
|
||||
use tokio::sync::Mutex;
|
||||
|
||||
lazy_static! {
|
||||
pub static ref ALPACA_MODE: Mode = env::var("ALPACA_MODE")
|
||||
.expect("ALPACA_MODE must be set.")
|
||||
.parse()
|
||||
.expect("ALPACA_MODE must be 'live' or 'paper'");
|
||||
static ref ALPACA_URL_SUBDOMAIN: String = match *ALPACA_MODE {
|
||||
Mode::Live => String::from("api"),
|
||||
Mode::Paper => String::from("paper-api"),
|
||||
};
|
||||
#[derive(Debug)]
|
||||
pub static ref ALPACA_API_URL: String = format!(
|
||||
"https://{subdomain}.alpaca.markets/v2",
|
||||
subdomain = *ALPACA_URL_SUBDOMAIN
|
||||
);
|
||||
#[derive(Debug)]
|
||||
pub static ref ALPACA_WEBSOCKET_URL: String = format!(
|
||||
"wss://{subdomain}.alpaca.markets/stream",
|
||||
subdomain = *ALPACA_URL_SUBDOMAIN
|
||||
);
|
||||
}
|
||||
|
||||
pub const ALPACA_STOCK_DATA_API_URL: &str = "https://data.alpaca.markets/v2/stocks/bars";
|
||||
pub const ALPACA_CRYPTO_DATA_API_URL: &str = "https://data.alpaca.markets/v1beta3/crypto/us/bars";
|
||||
pub const ALPACA_NEWS_DATA_API_URL: &str = "https://data.alpaca.markets/v1beta1/news";
|
||||
@@ -45,69 +24,78 @@ pub const ALPACA_CRYPTO_DATA_WEBSOCKET_URL: &str =
|
||||
"wss://stream.data.alpaca.markets/v1beta3/crypto/us";
|
||||
pub const ALPACA_NEWS_DATA_WEBSOCKET_URL: &str = "wss://stream.data.alpaca.markets/v1beta1/news";
|
||||
|
||||
lazy_static! {
|
||||
pub static ref ALPACA_MODE: Mode = env::var("ALPACA_MODE")
|
||||
.expect("ALPACA_MODE must be set.")
|
||||
.parse()
|
||||
.expect("ALPACA_MODE must be 'live' or 'paper'");
|
||||
pub static ref ALPACA_SOURCE: Source = env::var("ALPACA_SOURCE")
|
||||
.expect("ALPACA_SOURCE must be set.")
|
||||
.parse()
|
||||
.expect("ALPACA_SOURCE must be 'iex', 'sip', or 'otc'");
|
||||
pub static ref ALPACA_API_KEY: String = env::var("ALPACA_API_KEY").expect("ALPACA_API_KEY must be set.");
|
||||
pub static ref ALPACA_API_SECRET: String = env::var("ALPACA_API_SECRET").expect("ALPACA_API_SECRET must be set.");
|
||||
#[derive(Debug)]
|
||||
pub static ref ALPACA_API_URL: String = format!(
|
||||
"https://{}.alpaca.markets/v2",
|
||||
match *ALPACA_MODE {
|
||||
Mode::Live => String::from("api"),
|
||||
Mode::Paper => String::from("paper-api"),
|
||||
}
|
||||
);
|
||||
#[derive(Debug)]
|
||||
pub static ref ALPACA_WEBSOCKET_URL: String = format!(
|
||||
"wss://{}.alpaca.markets/stream",
|
||||
match *ALPACA_MODE {
|
||||
Mode::Live => String::from("api"),
|
||||
Mode::Paper => String::from("paper-api"),
|
||||
}
|
||||
);
|
||||
pub static ref MAX_BERT_INPUTS: usize = env::var("MAX_BERT_INPUTS")
|
||||
.expect("MAX_BERT_INPUTS must be set.")
|
||||
.parse()
|
||||
.expect("MAX_BERT_INPUTS must be a positive integer.");
|
||||
|
||||
}
|
||||
|
||||
pub struct Config {
|
||||
pub alpaca_api_key: String,
|
||||
pub alpaca_api_secret: String,
|
||||
pub alpaca_client: Client,
|
||||
pub alpaca_rate_limit: DefaultDirectRateLimiter,
|
||||
pub alpaca_source: Source,
|
||||
pub alpaca_rate_limiter: DefaultDirectRateLimiter,
|
||||
pub clickhouse_client: clickhouse::Client,
|
||||
pub max_bert_inputs: usize,
|
||||
pub sequence_classifier: Arc<Mutex<SequenceClassificationModel>>,
|
||||
pub sequence_classifier: Mutex<SequenceClassificationModel>,
|
||||
}
|
||||
|
||||
impl Config {
|
||||
pub fn from_env() -> Self {
|
||||
let alpaca_api_key = env::var("ALPACA_API_KEY").expect("ALPACA_API_KEY must be set.");
|
||||
let alpaca_api_secret =
|
||||
env::var("ALPACA_API_SECRET").expect("ALPACA_API_SECRET must be set.");
|
||||
let alpaca_source: Source = env::var("ALPACA_SOURCE")
|
||||
.expect("ALPACA_SOURCE must be set.")
|
||||
.parse()
|
||||
.expect("ALPACA_SOURCE must be 'iex', 'sip', or 'otc'.");
|
||||
|
||||
let clickhouse_url = env::var("CLICKHOUSE_URL").expect("CLICKHOUSE_URL must be set.");
|
||||
let clickhouse_user = env::var("CLICKHOUSE_USER").expect("CLICKHOUSE_USER must be set.");
|
||||
let clickhouse_password =
|
||||
env::var("CLICKHOUSE_PASSWORD").expect("CLICKHOUSE_PASSWORD must be set.");
|
||||
let clickhouse_db = env::var("CLICKHOUSE_DB").expect("CLICKHOUSE_DB must be set.");
|
||||
|
||||
let max_bert_inputs: usize = env::var("MAX_BERT_INPUTS")
|
||||
.expect("MAX_BERT_INPUTS must be set.")
|
||||
.parse()
|
||||
.expect("MAX_BERT_INPUTS must be a positive integer.");
|
||||
|
||||
Self {
|
||||
alpaca_client: Client::builder()
|
||||
.default_headers(HeaderMap::from_iter([
|
||||
(
|
||||
HeaderName::from_static("apca-api-key-id"),
|
||||
HeaderValue::from_str(&alpaca_api_key)
|
||||
HeaderValue::from_str(&ALPACA_API_KEY)
|
||||
.expect("Alpaca API key must not contain invalid characters."),
|
||||
),
|
||||
(
|
||||
HeaderName::from_static("apca-api-secret-key"),
|
||||
HeaderValue::from_str(&alpaca_api_secret)
|
||||
HeaderValue::from_str(&ALPACA_API_SECRET)
|
||||
.expect("Alpaca API secret must not contain invalid characters."),
|
||||
),
|
||||
]))
|
||||
.build()
|
||||
.unwrap(),
|
||||
alpaca_rate_limit: RateLimiter::direct(Quota::per_minute(match alpaca_source {
|
||||
alpaca_rate_limiter: RateLimiter::direct(Quota::per_minute(match *ALPACA_SOURCE {
|
||||
Source::Iex => unsafe { NonZeroU32::new_unchecked(200) },
|
||||
Source::Sip => unsafe { NonZeroU32::new_unchecked(10000) },
|
||||
Source::Otc => unimplemented!("OTC rate limit not implemented."),
|
||||
})),
|
||||
alpaca_source,
|
||||
clickhouse_client: clickhouse::Client::default()
|
||||
.with_url(clickhouse_url)
|
||||
.with_user(clickhouse_user)
|
||||
.with_password(clickhouse_password)
|
||||
.with_database(clickhouse_db),
|
||||
alpaca_api_key,
|
||||
alpaca_api_secret,
|
||||
max_bert_inputs,
|
||||
sequence_classifier: Arc::new(Mutex::new(
|
||||
.with_url(env::var("CLICKHOUSE_URL").expect("CLICKHOUSE_URL must be set."))
|
||||
.with_user(env::var("CLICKHOUSE_USER").expect("CLICKHOUSE_USER must be set."))
|
||||
.with_password(
|
||||
env::var("CLICKHOUSE_PASSWORD").expect("CLICKHOUSE_PASSWORD must be set."),
|
||||
)
|
||||
.with_database(env::var("CLICKHOUSE_DB").expect("CLICKHOUSE_DB must be set.")),
|
||||
sequence_classifier: Mutex::new(
|
||||
SequenceClassificationModel::new(SequenceClassificationConfig::new(
|
||||
ModelType::Bert,
|
||||
ModelResource::Torch(Box::new(LocalResource {
|
||||
@@ -125,7 +113,7 @@ impl Config {
|
||||
None,
|
||||
))
|
||||
.unwrap(),
|
||||
)),
|
||||
),
|
||||
}
|
||||
}
|
||||
|
||||
|
68
src/init.rs
68
src/init.rs
@@ -4,13 +4,18 @@ use crate::{
|
||||
types::alpaca::{self, api, shared::Sort},
|
||||
};
|
||||
use log::{info, warn};
|
||||
use std::{collections::HashSet, sync::Arc};
|
||||
use std::{collections::HashMap, sync::Arc};
|
||||
use time::OffsetDateTime;
|
||||
use tokio::join;
|
||||
|
||||
pub async fn check_account(config: &Arc<Config>) {
|
||||
let account = alpaca::api::incoming::account::get(config, None)
|
||||
.await
|
||||
.unwrap();
|
||||
let account = alpaca::api::incoming::account::get(
|
||||
&config.alpaca_client,
|
||||
&config.alpaca_rate_limiter,
|
||||
None,
|
||||
)
|
||||
.await
|
||||
.unwrap();
|
||||
|
||||
assert!(
|
||||
!(account.status != alpaca::api::incoming::account::Status::Active),
|
||||
@@ -41,7 +46,8 @@ pub async fn rehydrate_orders(config: &Arc<Config>) {
|
||||
let mut after = OffsetDateTime::UNIX_EPOCH;
|
||||
|
||||
while let Some(message) = api::incoming::order::get(
|
||||
config,
|
||||
&config.alpaca_client,
|
||||
&config.alpaca_rate_limiter,
|
||||
&api::outgoing::order::Order {
|
||||
status: Some(api::outgoing::order::Status::All),
|
||||
limit: Some(500),
|
||||
@@ -74,30 +80,52 @@ pub async fn rehydrate_orders(config: &Arc<Config>) {
|
||||
info!("Rehydrated order data.");
|
||||
}
|
||||
|
||||
pub async fn check_positions(config: &Arc<Config>) {
|
||||
pub async fn rehydrate_positions(config: &Arc<Config>) {
|
||||
info!("Rehydrating position data.");
|
||||
|
||||
let positions_future = async {
|
||||
alpaca::api::incoming::position::get(config, None)
|
||||
.await
|
||||
.unwrap()
|
||||
alpaca::api::incoming::position::get(
|
||||
&config.alpaca_client,
|
||||
&config.alpaca_rate_limiter,
|
||||
None,
|
||||
)
|
||||
.await
|
||||
.unwrap()
|
||||
.into_iter()
|
||||
.map(|position| (position.symbol.clone(), position))
|
||||
.collect::<HashMap<_, _>>()
|
||||
};
|
||||
|
||||
let assets_future = async {
|
||||
database::assets::select(&config.clickhouse_client)
|
||||
.await
|
||||
.unwrap()
|
||||
.into_iter()
|
||||
.map(|asset| asset.symbol)
|
||||
.collect::<HashSet<_>>()
|
||||
};
|
||||
|
||||
let (positions, assets) = tokio::join!(positions_future, assets_future);
|
||||
let (mut positions, assets) = join!(positions_future, assets_future);
|
||||
|
||||
for position in positions {
|
||||
if !assets.contains(&position.symbol) {
|
||||
warn!(
|
||||
"Position for unmonitored asset: {}, {} shares.",
|
||||
position.symbol, position.qty
|
||||
);
|
||||
}
|
||||
let assets = assets
|
||||
.into_iter()
|
||||
.map(|mut asset| {
|
||||
if let Some(position) = positions.remove(&asset.symbol) {
|
||||
asset.qty = position.qty_available;
|
||||
} else {
|
||||
asset.qty = 0.0;
|
||||
}
|
||||
asset
|
||||
})
|
||||
.collect::<Vec<_>>();
|
||||
|
||||
database::assets::upsert_batch(&config.clickhouse_client, &assets)
|
||||
.await
|
||||
.unwrap();
|
||||
|
||||
for position in positions.values() {
|
||||
warn!(
|
||||
"Position for unmonitored asset: {}, {} shares.",
|
||||
position.symbol, position.qty
|
||||
);
|
||||
}
|
||||
|
||||
info!("Rehydrated position data.");
|
||||
}
|
||||
|
22
src/main.rs
22
src/main.rs
@@ -13,7 +13,7 @@ mod utils;
|
||||
use config::Config;
|
||||
use dotenv::dotenv;
|
||||
use log4rs::config::Deserializers;
|
||||
use tokio::{spawn, sync::mpsc, try_join};
|
||||
use tokio::{join, spawn, sync::mpsc, try_join};
|
||||
|
||||
#[tokio::main]
|
||||
async fn main() {
|
||||
@@ -21,6 +21,12 @@ async fn main() {
|
||||
log4rs::init_file("log4rs.yaml", Deserializers::default()).unwrap();
|
||||
let config = Config::arc_from_env();
|
||||
|
||||
try_join!(
|
||||
database::backfills_bars::unfresh(&config.clickhouse_client),
|
||||
database::backfills_news::unfresh(&config.clickhouse_client)
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
database::cleanup_all(&config.clickhouse_client)
|
||||
.await
|
||||
.unwrap();
|
||||
@@ -28,15 +34,11 @@ async fn main() {
|
||||
.await
|
||||
.unwrap();
|
||||
|
||||
try_join!(
|
||||
database::backfills_bars::unfresh(&config.clickhouse_client),
|
||||
database::backfills_news::unfresh(&config.clickhouse_client)
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
init::check_account(&config).await;
|
||||
init::rehydrate_orders(&config).await;
|
||||
init::check_positions(&config).await;
|
||||
join!(
|
||||
init::rehydrate_orders(&config),
|
||||
init::rehydrate_positions(&config)
|
||||
);
|
||||
|
||||
spawn(threads::trading::run(config.clone()));
|
||||
|
||||
@@ -61,7 +63,7 @@ async fn main() {
|
||||
create_send_await!(
|
||||
data_sender,
|
||||
threads::data::Message::new,
|
||||
threads::data::Action::Add,
|
||||
threads::data::Action::Enable,
|
||||
assets
|
||||
);
|
||||
|
||||
|
@@ -50,14 +50,19 @@ pub async fn add(
|
||||
return Err(StatusCode::CONFLICT);
|
||||
}
|
||||
|
||||
let asset = alpaca::api::incoming::asset::get_by_symbol(&config, &request.symbol, None)
|
||||
.await
|
||||
.map_err(|e| {
|
||||
e.status()
|
||||
.map_or(StatusCode::INTERNAL_SERVER_ERROR, |status| {
|
||||
StatusCode::from_u16(status.as_u16()).unwrap()
|
||||
})
|
||||
})?;
|
||||
let asset = alpaca::api::incoming::asset::get_by_symbol(
|
||||
&config.alpaca_client,
|
||||
&config.alpaca_rate_limiter,
|
||||
&request.symbol,
|
||||
None,
|
||||
)
|
||||
.await
|
||||
.map_err(|e| {
|
||||
e.status()
|
||||
.map_or(StatusCode::INTERNAL_SERVER_ERROR, |status| {
|
||||
StatusCode::from_u16(status.as_u16()).unwrap()
|
||||
})
|
||||
})?;
|
||||
|
||||
if !asset.tradable || !asset.fractionable {
|
||||
return Err(StatusCode::FORBIDDEN);
|
||||
|
@@ -36,9 +36,13 @@ impl From<alpaca::api::incoming::clock::Clock> for Message {
|
||||
|
||||
pub async fn run(config: Arc<Config>, sender: mpsc::Sender<Message>) {
|
||||
loop {
|
||||
let clock = alpaca::api::incoming::clock::get(&config, Some(backoff::infinite()))
|
||||
.await
|
||||
.unwrap();
|
||||
let clock = alpaca::api::incoming::clock::get(
|
||||
&config.alpaca_client,
|
||||
&config.alpaca_rate_limiter,
|
||||
Some(backoff::infinite()),
|
||||
)
|
||||
.await
|
||||
.unwrap();
|
||||
|
||||
let sleep_until = duration_until(if clock.is_open {
|
||||
info!("Market is open, will close at {}.", clock.next_close);
|
||||
|
@@ -1,6 +1,9 @@
|
||||
use super::ThreadType;
|
||||
use crate::{
|
||||
config::{Config, ALPACA_CRYPTO_DATA_API_URL, ALPACA_STOCK_DATA_API_URL},
|
||||
config::{
|
||||
Config, ALPACA_CRYPTO_DATA_API_URL, ALPACA_SOURCE, ALPACA_STOCK_DATA_API_URL,
|
||||
MAX_BERT_INPUTS,
|
||||
},
|
||||
database,
|
||||
types::{
|
||||
alpaca::{
|
||||
@@ -30,23 +33,24 @@ pub enum Action {
|
||||
Purge,
|
||||
}
|
||||
|
||||
impl From<super::Action> for Action {
|
||||
impl From<super::Action> for Option<Action> {
|
||||
fn from(action: super::Action) -> Self {
|
||||
match action {
|
||||
super::Action::Add => Self::Backfill,
|
||||
super::Action::Remove => Self::Purge,
|
||||
super::Action::Add | super::Action::Enable => Some(Action::Backfill),
|
||||
super::Action::Remove => Some(Action::Purge),
|
||||
super::Action::Disable => None,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub struct Message {
|
||||
pub action: Action,
|
||||
pub action: Option<Action>,
|
||||
pub symbols: Vec<String>,
|
||||
pub response: oneshot::Sender<()>,
|
||||
}
|
||||
|
||||
impl Message {
|
||||
pub fn new(action: Action, symbols: Vec<String>) -> (Self, oneshot::Receiver<()>) {
|
||||
pub fn new(action: Option<Action>, symbols: Vec<String>) -> (Self, oneshot::Receiver<()>) {
|
||||
let (sender, receiver) = oneshot::channel::<()>();
|
||||
(
|
||||
Self {
|
||||
@@ -77,7 +81,6 @@ pub async fn run(handler: Arc<Box<dyn Handler>>, mut receiver: mpsc::Receiver<Me
|
||||
|
||||
loop {
|
||||
let message = receiver.recv().await.unwrap();
|
||||
|
||||
spawn(handle_backfill_message(
|
||||
handler.clone(),
|
||||
backfill_jobs.clone(),
|
||||
@@ -94,7 +97,7 @@ async fn handle_backfill_message(
|
||||
let mut backfill_jobs = backfill_jobs.lock().await;
|
||||
|
||||
match message.action {
|
||||
Action::Backfill => {
|
||||
Some(Action::Backfill) => {
|
||||
let log_string = handler.log_string();
|
||||
|
||||
for symbol in message.symbols {
|
||||
@@ -134,7 +137,7 @@ async fn handle_backfill_message(
|
||||
);
|
||||
}
|
||||
}
|
||||
Action::Purge => {
|
||||
Some(Action::Purge) => {
|
||||
for symbol in &message.symbols {
|
||||
if let Some(job) = backfill_jobs.remove(symbol) {
|
||||
if !job.is_finished() {
|
||||
@@ -150,6 +153,7 @@ async fn handle_backfill_message(
|
||||
)
|
||||
.unwrap();
|
||||
}
|
||||
None => {}
|
||||
}
|
||||
|
||||
message.response.send(()).unwrap();
|
||||
@@ -159,7 +163,6 @@ struct BarHandler {
|
||||
config: Arc<Config>,
|
||||
data_url: &'static str,
|
||||
api_query_constructor: fn(
|
||||
config: &Arc<Config>,
|
||||
symbol: String,
|
||||
fetch_from: OffsetDateTime,
|
||||
fetch_to: OffsetDateTime,
|
||||
@@ -168,7 +171,6 @@ struct BarHandler {
|
||||
}
|
||||
|
||||
fn us_equity_query_constructor(
|
||||
config: &Arc<Config>,
|
||||
symbol: String,
|
||||
fetch_from: OffsetDateTime,
|
||||
fetch_to: OffsetDateTime,
|
||||
@@ -182,7 +184,7 @@ fn us_equity_query_constructor(
|
||||
limit: Some(10000),
|
||||
adjustment: None,
|
||||
asof: None,
|
||||
feed: Some(config.alpaca_source),
|
||||
feed: Some(*ALPACA_SOURCE),
|
||||
currency: None,
|
||||
page_token: next_page_token,
|
||||
sort: Some(Sort::Asc),
|
||||
@@ -190,7 +192,6 @@ fn us_equity_query_constructor(
|
||||
}
|
||||
|
||||
fn crypto_query_constructor(
|
||||
_: &Arc<Config>,
|
||||
symbol: String,
|
||||
fetch_from: OffsetDateTime,
|
||||
fetch_to: OffsetDateTime,
|
||||
@@ -226,7 +227,7 @@ impl Handler for BarHandler {
|
||||
}
|
||||
|
||||
async fn queue_backfill(&self, symbol: &str, fetch_to: OffsetDateTime) {
|
||||
if self.config.alpaca_source == Source::Iex {
|
||||
if *ALPACA_SOURCE == Source::Iex {
|
||||
let run_delay = duration_until(fetch_to + FIFTEEN_MINUTES + ONE_MINUTE);
|
||||
info!("Queing bar backfill for {} in {:?}.", symbol, run_delay);
|
||||
sleep(run_delay).await;
|
||||
@@ -241,10 +242,10 @@ impl Handler for BarHandler {
|
||||
|
||||
loop {
|
||||
let Ok(message) = api::incoming::bar::get_historical(
|
||||
&self.config,
|
||||
&self.config.alpaca_client,
|
||||
&self.config.alpaca_rate_limiter,
|
||||
self.data_url,
|
||||
&(self.api_query_constructor)(
|
||||
&self.config,
|
||||
symbol.clone(),
|
||||
fetch_from,
|
||||
fetch_to,
|
||||
@@ -328,7 +329,8 @@ impl Handler for NewsHandler {
|
||||
|
||||
loop {
|
||||
let Ok(message) = api::incoming::news::get_historical(
|
||||
&self.config,
|
||||
&self.config.alpaca_client,
|
||||
&self.config.alpaca_rate_limiter,
|
||||
&api::outgoing::news::News {
|
||||
symbols: vec![symbol.clone()],
|
||||
start: Some(fetch_from),
|
||||
@@ -367,18 +369,15 @@ impl Handler for NewsHandler {
|
||||
.map(|news| format!("{}\n\n{}", news.headline, news.content))
|
||||
.collect::<Vec<_>>();
|
||||
|
||||
let predictions = join_all(inputs.chunks(self.config.max_bert_inputs).map(|inputs| {
|
||||
let sequence_classifier = self.config.sequence_classifier.clone();
|
||||
async move {
|
||||
let sequence_classifier = sequence_classifier.lock().await;
|
||||
block_in_place(|| {
|
||||
sequence_classifier
|
||||
.predict(inputs.iter().map(String::as_str).collect::<Vec<_>>())
|
||||
.into_iter()
|
||||
.map(|label| Prediction::try_from(label).unwrap())
|
||||
.collect::<Vec<_>>()
|
||||
})
|
||||
}
|
||||
let predictions = join_all(inputs.chunks(*MAX_BERT_INPUTS).map(|inputs| async move {
|
||||
let sequence_classifier = self.config.sequence_classifier.lock().await;
|
||||
block_in_place(|| {
|
||||
sequence_classifier
|
||||
.predict(inputs.iter().map(String::as_str).collect::<Vec<_>>())
|
||||
.into_iter()
|
||||
.map(|label| Prediction::try_from(label).unwrap())
|
||||
.collect::<Vec<_>>()
|
||||
})
|
||||
}))
|
||||
.await
|
||||
.into_iter()
|
||||
|
@@ -4,7 +4,7 @@ mod websocket;
|
||||
use super::clock;
|
||||
use crate::{
|
||||
config::{
|
||||
Config, ALPACA_CRYPTO_DATA_WEBSOCKET_URL, ALPACA_NEWS_DATA_WEBSOCKET_URL,
|
||||
Config, ALPACA_CRYPTO_DATA_WEBSOCKET_URL, ALPACA_NEWS_DATA_WEBSOCKET_URL, ALPACA_SOURCE,
|
||||
ALPACA_STOCK_DATA_WEBSOCKET_URL,
|
||||
},
|
||||
create_send_await, database,
|
||||
@@ -21,9 +21,12 @@ use tokio::{
|
||||
use tokio_tungstenite::connect_async;
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
#[allow(dead_code)]
|
||||
pub enum Action {
|
||||
Add,
|
||||
Enable,
|
||||
Remove,
|
||||
Disable,
|
||||
}
|
||||
|
||||
pub struct Message {
|
||||
@@ -100,10 +103,7 @@ async fn init_thread(
|
||||
) {
|
||||
let websocket_url = match thread_type {
|
||||
ThreadType::Bars(Class::UsEquity) => {
|
||||
format!(
|
||||
"{}/{}",
|
||||
ALPACA_STOCK_DATA_WEBSOCKET_URL, &config.alpaca_source
|
||||
)
|
||||
format!("{}/{}", ALPACA_STOCK_DATA_WEBSOCKET_URL, *ALPACA_SOURCE)
|
||||
}
|
||||
ThreadType::Bars(Class::Crypto) => ALPACA_CRYPTO_DATA_WEBSOCKET_URL.into(),
|
||||
ThreadType::News => ALPACA_NEWS_DATA_WEBSOCKET_URL.into(),
|
||||
@@ -111,8 +111,7 @@ async fn init_thread(
|
||||
|
||||
let (websocket, _) = connect_async(websocket_url).await.unwrap();
|
||||
let (mut websocket_sink, mut websocket_stream) = websocket.split();
|
||||
alpaca::websocket::data::authenticate(&config, &mut websocket_sink, &mut websocket_stream)
|
||||
.await;
|
||||
alpaca::websocket::data::authenticate(&mut websocket_sink, &mut websocket_stream).await;
|
||||
|
||||
let (backfill_sender, backfill_receiver) = mpsc::channel(100);
|
||||
spawn(backfill::run(
|
||||
@@ -223,7 +222,8 @@ async fn handle_message(
|
||||
async move {
|
||||
let asset_future = async {
|
||||
alpaca::api::incoming::asset::get_by_symbol(
|
||||
&config,
|
||||
&config.alpaca_client,
|
||||
&config.alpaca_rate_limiter,
|
||||
&symbol,
|
||||
Some(backoff::infinite()),
|
||||
)
|
||||
@@ -233,7 +233,8 @@ async fn handle_message(
|
||||
|
||||
let position_future = async {
|
||||
alpaca::api::incoming::position::get_by_symbol(
|
||||
&config,
|
||||
&config.alpaca_rate_limiter,
|
||||
&config.alpaca_client,
|
||||
&symbol,
|
||||
Some(backoff::infinite()),
|
||||
)
|
||||
@@ -256,6 +257,7 @@ async fn handle_message(
|
||||
.await
|
||||
.unwrap();
|
||||
}
|
||||
_ => {}
|
||||
}
|
||||
|
||||
message.response.send(()).unwrap();
|
||||
@@ -292,7 +294,7 @@ async fn handle_clock_message(
|
||||
create_send_await!(
|
||||
bars_us_equity_backfill_sender,
|
||||
backfill::Message::new,
|
||||
backfill::Action::Backfill,
|
||||
Some(backfill::Action::Backfill),
|
||||
us_equity_symbols.clone()
|
||||
);
|
||||
};
|
||||
@@ -301,7 +303,7 @@ async fn handle_clock_message(
|
||||
create_send_await!(
|
||||
bars_crypto_backfill_sender,
|
||||
backfill::Message::new,
|
||||
backfill::Action::Backfill,
|
||||
Some(backfill::Action::Backfill),
|
||||
crypto_symbols.clone()
|
||||
);
|
||||
};
|
||||
@@ -310,7 +312,7 @@ async fn handle_clock_message(
|
||||
create_send_await!(
|
||||
news_backfill_sender,
|
||||
backfill::Message::new,
|
||||
backfill::Action::Backfill,
|
||||
Some(backfill::Action::Backfill),
|
||||
symbols
|
||||
);
|
||||
};
|
||||
|
@@ -26,23 +26,23 @@ pub enum Action {
|
||||
Unsubscribe,
|
||||
}
|
||||
|
||||
impl From<super::Action> for Action {
|
||||
impl From<super::Action> for Option<Action> {
|
||||
fn from(action: super::Action) -> Self {
|
||||
match action {
|
||||
super::Action::Add => Self::Subscribe,
|
||||
super::Action::Remove => Self::Unsubscribe,
|
||||
super::Action::Add | super::Action::Enable => Some(Action::Subscribe),
|
||||
super::Action::Remove | super::Action::Disable => Some(Action::Unsubscribe),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub struct Message {
|
||||
pub action: Action,
|
||||
pub action: Option<Action>,
|
||||
pub symbols: Vec<String>,
|
||||
pub response: oneshot::Sender<()>,
|
||||
}
|
||||
|
||||
impl Message {
|
||||
pub fn new(action: Action, symbols: Vec<String>) -> (Self, oneshot::Receiver<()>) {
|
||||
pub fn new(action: Option<Action>, symbols: Vec<String>) -> (Self, oneshot::Receiver<()>) {
|
||||
let (sender, receiver) = oneshot::channel();
|
||||
(
|
||||
Self {
|
||||
@@ -115,7 +115,7 @@ async fn handle_message(
|
||||
message: Message,
|
||||
) {
|
||||
match message.action {
|
||||
Action::Subscribe => {
|
||||
Some(Action::Subscribe) => {
|
||||
let (pending_subscriptions, receivers): (Vec<_>, Vec<_>) = message
|
||||
.symbols
|
||||
.iter()
|
||||
@@ -144,7 +144,7 @@ async fn handle_message(
|
||||
|
||||
join_all(receivers).await;
|
||||
}
|
||||
Action::Unsubscribe => {
|
||||
Some(Action::Unsubscribe) => {
|
||||
let (pending_unsubscriptions, receivers): (Vec<_>, Vec<_>) = message
|
||||
.symbols
|
||||
.iter()
|
||||
@@ -173,6 +173,7 @@ async fn handle_message(
|
||||
|
||||
join_all(receivers).await;
|
||||
}
|
||||
None => {}
|
||||
}
|
||||
|
||||
message.response.send(()).unwrap();
|
||||
|
@@ -13,8 +13,7 @@ pub async fn run(config: Arc<Config>) {
|
||||
let (websocket, _) = connect_async(&*ALPACA_WEBSOCKET_URL).await.unwrap();
|
||||
let (mut websocket_sink, mut websocket_stream) = websocket.split();
|
||||
|
||||
alpaca::websocket::trading::authenticate(&config, &mut websocket_sink, &mut websocket_stream)
|
||||
.await;
|
||||
alpaca::websocket::trading::authenticate(&mut websocket_sink, &mut websocket_stream).await;
|
||||
alpaca::websocket::trading::subscribe(&mut websocket_sink, &mut websocket_stream).await;
|
||||
|
||||
spawn(websocket::run(config, websocket_stream, websocket_sink));
|
||||
|
@@ -22,10 +22,8 @@ pub async fn run(
|
||||
|
||||
loop {
|
||||
let message = websocket_stream.next().await.unwrap().unwrap();
|
||||
let config = config.clone();
|
||||
|
||||
spawn(handle_websocket_message(
|
||||
config,
|
||||
config.clone(),
|
||||
websocket_sink.clone(),
|
||||
message,
|
||||
));
|
||||
@@ -42,7 +40,7 @@ async fn handle_websocket_message(
|
||||
if let Ok(message) = from_str::<websocket::trading::incoming::Message>(
|
||||
&String::from_utf8_lossy(&message),
|
||||
) {
|
||||
spawn(handle_parsed_websocket_message(config.clone(), message));
|
||||
handle_parsed_websocket_message(config, message).await;
|
||||
} else {
|
||||
error!("Failed to deserialize websocket message: {:?}", message);
|
||||
}
|
||||
|
@@ -1,12 +1,13 @@
|
||||
use crate::config::{Config, ALPACA_API_URL};
|
||||
use crate::config::ALPACA_API_URL;
|
||||
use backoff::{future::retry_notify, ExponentialBackoff};
|
||||
use governor::DefaultDirectRateLimiter;
|
||||
use log::warn;
|
||||
use reqwest::Error;
|
||||
use reqwest::{Client, Error};
|
||||
use serde::Deserialize;
|
||||
use serde_aux::field_attributes::{
|
||||
deserialize_number_from_string, deserialize_option_number_from_string,
|
||||
};
|
||||
use std::{sync::Arc, time::Duration};
|
||||
use std::time::Duration;
|
||||
use time::OffsetDateTime;
|
||||
use uuid::Uuid;
|
||||
|
||||
@@ -80,15 +81,15 @@ pub struct Account {
|
||||
}
|
||||
|
||||
pub async fn get(
|
||||
config: &Arc<Config>,
|
||||
alpaca_client: &Client,
|
||||
alpaca_rate_limiter: &DefaultDirectRateLimiter,
|
||||
backoff: Option<ExponentialBackoff>,
|
||||
) -> Result<Account, Error> {
|
||||
retry_notify(
|
||||
backoff.unwrap_or_default(),
|
||||
|| async {
|
||||
config.alpaca_rate_limit.until_ready().await;
|
||||
config
|
||||
.alpaca_client
|
||||
alpaca_rate_limiter.until_ready().await;
|
||||
alpaca_client
|
||||
.get(&format!("{}/account", *ALPACA_API_URL))
|
||||
.send()
|
||||
.await?
|
||||
|
@@ -1,17 +1,18 @@
|
||||
use super::position::Position;
|
||||
use crate::{
|
||||
config::{Config, ALPACA_API_URL},
|
||||
config::ALPACA_API_URL,
|
||||
types::{
|
||||
self,
|
||||
alpaca::shared::asset::{Class, Exchange, Status},
|
||||
},
|
||||
};
|
||||
use backoff::{future::retry_notify, ExponentialBackoff};
|
||||
use governor::DefaultDirectRateLimiter;
|
||||
use log::warn;
|
||||
use reqwest::Error;
|
||||
use reqwest::{Client, Error};
|
||||
use serde::Deserialize;
|
||||
use serde_aux::field_attributes::deserialize_option_number_from_string;
|
||||
use std::{sync::Arc, time::Duration};
|
||||
use std::time::Duration;
|
||||
use uuid::Uuid;
|
||||
|
||||
#[allow(clippy::struct_excessive_bools)]
|
||||
@@ -47,16 +48,16 @@ impl From<(Asset, Option<Position>)> for types::Asset {
|
||||
}
|
||||
|
||||
pub async fn get_by_symbol(
|
||||
config: &Arc<Config>,
|
||||
alpaca_client: &Client,
|
||||
alpaca_rate_limiter: &DefaultDirectRateLimiter,
|
||||
symbol: &str,
|
||||
backoff: Option<ExponentialBackoff>,
|
||||
) -> Result<Asset, Error> {
|
||||
retry_notify(
|
||||
backoff.unwrap_or_default(),
|
||||
|| async {
|
||||
config.alpaca_rate_limit.until_ready().await;
|
||||
config
|
||||
.alpaca_client
|
||||
alpaca_rate_limiter.until_ready().await;
|
||||
alpaca_client
|
||||
.get(&format!("{}/assets/{}", *ALPACA_API_URL, symbol))
|
||||
.send()
|
||||
.await?
|
||||
|
@@ -1,12 +1,10 @@
|
||||
use crate::{
|
||||
config::Config,
|
||||
types::{self, alpaca::api::outgoing},
|
||||
};
|
||||
use crate::types::{self, alpaca::api::outgoing};
|
||||
use backoff::{future::retry_notify, ExponentialBackoff};
|
||||
use governor::DefaultDirectRateLimiter;
|
||||
use log::warn;
|
||||
use reqwest::Error;
|
||||
use reqwest::{Client, Error};
|
||||
use serde::Deserialize;
|
||||
use std::{collections::HashMap, sync::Arc, time::Duration};
|
||||
use std::{collections::HashMap, time::Duration};
|
||||
use time::OffsetDateTime;
|
||||
|
||||
#[derive(Deserialize)]
|
||||
@@ -53,7 +51,8 @@ pub struct Message {
|
||||
}
|
||||
|
||||
pub async fn get_historical(
|
||||
config: &Arc<Config>,
|
||||
alpaca_client: &Client,
|
||||
alpaca_rate_limiter: &DefaultDirectRateLimiter,
|
||||
data_url: &str,
|
||||
query: &outgoing::bar::Bar,
|
||||
backoff: Option<ExponentialBackoff>,
|
||||
@@ -61,9 +60,8 @@ pub async fn get_historical(
|
||||
retry_notify(
|
||||
backoff.unwrap_or_default(),
|
||||
|| async {
|
||||
config.alpaca_rate_limit.until_ready().await;
|
||||
config
|
||||
.alpaca_client
|
||||
alpaca_rate_limiter.until_ready().await;
|
||||
alpaca_client
|
||||
.get(data_url)
|
||||
.query(query)
|
||||
.send()
|
||||
|
@@ -1,9 +1,10 @@
|
||||
use crate::config::{Config, ALPACA_API_URL};
|
||||
use crate::config::ALPACA_API_URL;
|
||||
use backoff::{future::retry_notify, ExponentialBackoff};
|
||||
use governor::DefaultDirectRateLimiter;
|
||||
use log::warn;
|
||||
use reqwest::Error;
|
||||
use reqwest::{Client, Error};
|
||||
use serde::Deserialize;
|
||||
use std::{sync::Arc, time::Duration};
|
||||
use std::time::Duration;
|
||||
use time::OffsetDateTime;
|
||||
|
||||
#[derive(Deserialize)]
|
||||
@@ -18,15 +19,15 @@ pub struct Clock {
|
||||
}
|
||||
|
||||
pub async fn get(
|
||||
config: &Arc<Config>,
|
||||
alpaca_client: &Client,
|
||||
alpaca_rate_limiter: &DefaultDirectRateLimiter,
|
||||
backoff: Option<ExponentialBackoff>,
|
||||
) -> Result<Clock, Error> {
|
||||
retry_notify(
|
||||
backoff.unwrap_or_default(),
|
||||
|| async {
|
||||
config.alpaca_rate_limit.until_ready().await;
|
||||
config
|
||||
.alpaca_client
|
||||
alpaca_rate_limiter.until_ready().await;
|
||||
alpaca_client
|
||||
.get(&format!("{}/clock", *ALPACA_API_URL))
|
||||
.send()
|
||||
.await?
|
||||
|
@@ -1,5 +1,5 @@
|
||||
use crate::{
|
||||
config::{Config, ALPACA_NEWS_DATA_API_URL},
|
||||
config::ALPACA_NEWS_DATA_API_URL,
|
||||
types::{
|
||||
self,
|
||||
alpaca::{api::outgoing, shared::news::normalize_html_content},
|
||||
@@ -7,10 +7,11 @@ use crate::{
|
||||
utils::de,
|
||||
};
|
||||
use backoff::{future::retry_notify, ExponentialBackoff};
|
||||
use governor::DefaultDirectRateLimiter;
|
||||
use log::warn;
|
||||
use reqwest::Error;
|
||||
use reqwest::{Client, Error};
|
||||
use serde::Deserialize;
|
||||
use std::{sync::Arc, time::Duration};
|
||||
use std::time::Duration;
|
||||
use time::OffsetDateTime;
|
||||
|
||||
#[derive(Deserialize)]
|
||||
@@ -73,16 +74,16 @@ pub struct Message {
|
||||
}
|
||||
|
||||
pub async fn get_historical(
|
||||
config: &Arc<Config>,
|
||||
alpaca_client: &Client,
|
||||
alpaca_rate_limiter: &DefaultDirectRateLimiter,
|
||||
query: &outgoing::news::News,
|
||||
backoff: Option<ExponentialBackoff>,
|
||||
) -> Result<Message, Error> {
|
||||
retry_notify(
|
||||
backoff.unwrap_or_default(),
|
||||
|| async {
|
||||
config.alpaca_rate_limit.until_ready().await;
|
||||
config
|
||||
.alpaca_client
|
||||
alpaca_rate_limiter.until_ready().await;
|
||||
alpaca_client
|
||||
.get(ALPACA_NEWS_DATA_API_URL)
|
||||
.query(query)
|
||||
.send()
|
||||
|
@@ -1,25 +1,26 @@
|
||||
use crate::{
|
||||
config::{Config, ALPACA_API_URL},
|
||||
config::ALPACA_API_URL,
|
||||
types::alpaca::{api::outgoing, shared},
|
||||
};
|
||||
use backoff::{future::retry_notify, ExponentialBackoff};
|
||||
use governor::DefaultDirectRateLimiter;
|
||||
use log::warn;
|
||||
use reqwest::Error;
|
||||
use std::{sync::Arc, time::Duration};
|
||||
use reqwest::{Client, Error};
|
||||
use std::time::Duration;
|
||||
|
||||
pub use shared::order::Order;
|
||||
|
||||
pub async fn get(
|
||||
config: &Arc<Config>,
|
||||
alpaca_client: &Client,
|
||||
alpaca_rate_limiter: &DefaultDirectRateLimiter,
|
||||
query: &outgoing::order::Order,
|
||||
backoff: Option<ExponentialBackoff>,
|
||||
) -> Result<Vec<Order>, Error> {
|
||||
retry_notify(
|
||||
backoff.unwrap_or_default(),
|
||||
|| async {
|
||||
config.alpaca_rate_limit.until_ready().await;
|
||||
config
|
||||
.alpaca_client
|
||||
alpaca_rate_limiter.until_ready().await;
|
||||
alpaca_client
|
||||
.get(&format!("{}/orders", *ALPACA_API_URL))
|
||||
.query(query)
|
||||
.send()
|
||||
|
@@ -1,5 +1,5 @@
|
||||
use crate::{
|
||||
config::{Config, ALPACA_API_URL},
|
||||
config::ALPACA_API_URL,
|
||||
types::alpaca::shared::{
|
||||
self,
|
||||
asset::{Class, Exchange},
|
||||
@@ -7,10 +7,12 @@ use crate::{
|
||||
utils::de,
|
||||
};
|
||||
use backoff::{future::retry_notify, ExponentialBackoff};
|
||||
use governor::DefaultDirectRateLimiter;
|
||||
use log::warn;
|
||||
use reqwest::Client;
|
||||
use serde::Deserialize;
|
||||
use serde_aux::field_attributes::deserialize_number_from_string;
|
||||
use std::{sync::Arc, time::Duration};
|
||||
use std::time::Duration;
|
||||
use uuid::Uuid;
|
||||
|
||||
#[derive(Deserialize)]
|
||||
@@ -65,15 +67,15 @@ pub struct Position {
|
||||
}
|
||||
|
||||
pub async fn get(
|
||||
config: &Arc<Config>,
|
||||
alpaca_client: &Client,
|
||||
alpaca_rate_limiter: &DefaultDirectRateLimiter,
|
||||
backoff: Option<ExponentialBackoff>,
|
||||
) -> Result<Vec<Position>, reqwest::Error> {
|
||||
retry_notify(
|
||||
backoff.unwrap_or_default(),
|
||||
|| async {
|
||||
config.alpaca_rate_limit.until_ready().await;
|
||||
config
|
||||
.alpaca_client
|
||||
alpaca_rate_limiter.until_ready().await;
|
||||
alpaca_client
|
||||
.get(&format!("{}/positions", *ALPACA_API_URL))
|
||||
.send()
|
||||
.await?
|
||||
@@ -98,16 +100,16 @@ pub async fn get(
|
||||
}
|
||||
|
||||
pub async fn get_by_symbol(
|
||||
config: &Arc<Config>,
|
||||
alpaca_rate_limiter: &DefaultDirectRateLimiter,
|
||||
alpaca_client: &Client,
|
||||
symbol: &str,
|
||||
backoff: Option<ExponentialBackoff>,
|
||||
) -> Result<Option<Position>, reqwest::Error> {
|
||||
retry_notify(
|
||||
backoff.unwrap_or_default(),
|
||||
|| async {
|
||||
config.alpaca_rate_limit.until_ready().await;
|
||||
let response = config
|
||||
.alpaca_client
|
||||
alpaca_rate_limiter.until_ready().await;
|
||||
let response = alpaca_client
|
||||
.get(&format!("{}/positions/{}", *ALPACA_API_URL, symbol))
|
||||
.send()
|
||||
.await?;
|
||||
|
@@ -1,19 +1,20 @@
|
||||
pub mod incoming;
|
||||
pub mod outgoing;
|
||||
|
||||
use crate::{config::Config, types::alpaca::websocket};
|
||||
use crate::{
|
||||
config::{ALPACA_API_KEY, ALPACA_API_SECRET},
|
||||
types::alpaca::websocket,
|
||||
};
|
||||
use core::panic;
|
||||
use futures_util::{
|
||||
stream::{SplitSink, SplitStream},
|
||||
SinkExt, StreamExt,
|
||||
};
|
||||
use serde_json::{from_str, to_string};
|
||||
use std::sync::Arc;
|
||||
use tokio::net::TcpStream;
|
||||
use tokio_tungstenite::{tungstenite::Message, MaybeTlsStream, WebSocketStream};
|
||||
|
||||
pub async fn authenticate(
|
||||
config: &Arc<Config>,
|
||||
sink: &mut SplitSink<WebSocketStream<MaybeTlsStream<TcpStream>>, Message>,
|
||||
stream: &mut SplitStream<WebSocketStream<MaybeTlsStream<TcpStream>>>,
|
||||
) {
|
||||
@@ -31,8 +32,8 @@ pub async fn authenticate(
|
||||
sink.send(Message::Text(
|
||||
to_string(&websocket::data::outgoing::Message::Auth(
|
||||
websocket::auth::Message {
|
||||
key: config.alpaca_api_key.clone(),
|
||||
secret: config.alpaca_api_secret.clone(),
|
||||
key: (*ALPACA_API_KEY).clone(),
|
||||
secret: (*ALPACA_API_SECRET).clone(),
|
||||
},
|
||||
))
|
||||
.unwrap(),
|
||||
|
@@ -1,27 +1,28 @@
|
||||
pub mod incoming;
|
||||
pub mod outgoing;
|
||||
|
||||
use crate::{config::Config, types::alpaca::websocket};
|
||||
use crate::{
|
||||
config::{ALPACA_API_KEY, ALPACA_API_SECRET},
|
||||
types::alpaca::websocket,
|
||||
};
|
||||
use core::panic;
|
||||
use futures_util::{
|
||||
stream::{SplitSink, SplitStream},
|
||||
SinkExt, StreamExt,
|
||||
};
|
||||
use serde_json::{from_str, to_string};
|
||||
use std::sync::Arc;
|
||||
use tokio::net::TcpStream;
|
||||
use tokio_tungstenite::{tungstenite::Message, MaybeTlsStream, WebSocketStream};
|
||||
|
||||
pub async fn authenticate(
|
||||
config: &Arc<Config>,
|
||||
sink: &mut SplitSink<WebSocketStream<MaybeTlsStream<TcpStream>>, Message>,
|
||||
stream: &mut SplitStream<WebSocketStream<MaybeTlsStream<TcpStream>>>,
|
||||
) {
|
||||
sink.send(Message::Text(
|
||||
to_string(&websocket::trading::outgoing::Message::Auth(
|
||||
websocket::auth::Message {
|
||||
key: config.alpaca_api_key.clone(),
|
||||
secret: config.alpaca_api_secret.clone(),
|
||||
key: (*ALPACA_API_KEY).clone(),
|
||||
secret: (*ALPACA_API_SECRET).clone(),
|
||||
},
|
||||
))
|
||||
.unwrap(),
|
||||
|
Reference in New Issue
Block a user