web3-proxy/web3_proxy/src/rpcs/request.rs

405 lines
15 KiB
Rust
Raw Normal View History

use super::one::Web3Rpc;
2023-06-25 01:24:46 +03:00
use crate::errors::{Web3ProxyErrorContext, Web3ProxyResult};
2023-06-13 21:51:19 +03:00
use crate::frontend::authorization::{Authorization, AuthorizationType};
use crate::jsonrpc::{JsonRpcParams, JsonRpcResultData};
use anyhow::Context;
use chrono::Utc;
use derive_more::From;
2022-11-01 21:54:39 +03:00
use entities::revert_log;
use entities::sea_orm_active_enums::Method;
2023-04-11 01:01:36 +03:00
use ethers::providers::ProviderError;
use ethers::types::{Address, Bytes};
2022-11-14 21:24:52 +03:00
use migration::sea_orm::{self, ActiveEnum, ActiveModelTrait};
2023-06-13 05:13:06 +03:00
use nanorand::Rng;
use serde_json::json;
use std::sync::atomic;
2022-08-24 03:14:49 +03:00
use std::sync::Arc;
2023-05-24 00:40:34 +03:00
use tokio::time::{Duration, Instant};
use tracing::{debug, error, trace, warn, Level};
2022-08-24 03:11:49 +03:00
#[derive(Debug, From)]
2022-08-24 03:14:49 +03:00
pub enum OpenRequestResult {
2022-08-24 03:59:05 +03:00
Handle(OpenRequestHandle),
/// Unable to start a request. Retry at the given time.
2022-08-24 03:11:49 +03:00
RetryAt(Instant),
2023-02-15 04:41:40 +03:00
/// Unable to start a request because no servers are synced
NotReady,
2022-08-24 03:11:49 +03:00
}
2022-08-24 03:59:05 +03:00
/// Make RPC requests through this handle and drop it when you are done.
/// Opening this handle checks rate limits. Developers, try to keep opening a handle and using it as close together as possible
2022-08-30 23:01:42 +03:00
#[derive(Debug)]
2022-09-09 06:53:16 +03:00
pub struct OpenRequestHandle {
authorization: Arc<Authorization>,
error_handler: RequestErrorHandler,
2023-02-12 21:22:20 +03:00
rpc: Arc<Web3Rpc>,
2022-09-09 06:53:16 +03:00
}
2022-08-24 03:11:49 +03:00
2023-04-06 01:34:28 +03:00
/// Depending on the context, RPC errors require different handling.
2023-06-25 06:39:46 +03:00
#[derive(Copy, Clone, Debug, Default)]
2023-04-06 01:34:28 +03:00
pub enum RequestErrorHandler {
/// Log at the trace level. Use when errors are expected.
2023-06-25 06:39:46 +03:00
#[default]
TraceLevel,
2022-09-23 01:15:56 +03:00
/// Log at the debug level. Use when errors are expected.
DebugLevel,
2022-09-23 01:15:56 +03:00
/// Log at the error level. Use when errors are bad.
ErrorLevel,
2022-09-23 01:15:56 +03:00
/// Log at the warn level. Use when errors do not cause problems.
WarnLevel,
/// Potentially save the revert. Users can tune how often this happens
2023-04-04 15:40:22 +03:00
Save,
}
2022-09-24 10:04:11 +03:00
// TODO: second param could be skipped since we don't need it here
#[derive(serde::Deserialize, serde::Serialize)]
2022-09-24 10:04:11 +03:00
struct EthCallParams((EthCallFirstParams, Option<serde_json::Value>));
#[derive(serde::Deserialize, serde::Serialize)]
struct EthCallFirstParams {
to: Address,
data: Option<Bytes>,
}
2023-04-06 01:34:28 +03:00
impl From<Level> for RequestErrorHandler {
fn from(level: Level) -> Self {
match level {
Level::TRACE => RequestErrorHandler::TraceLevel,
Level::DEBUG => RequestErrorHandler::DebugLevel,
Level::ERROR => RequestErrorHandler::ErrorLevel,
Level::WARN => RequestErrorHandler::WarnLevel,
2022-09-23 01:15:56 +03:00
_ => unimplemented!("unexpected tracing Level"),
}
}
}
impl Authorization {
2022-09-23 01:34:43 +03:00
/// Save a RPC call that return "execution reverted" to the database.
async fn save_revert(
self: Arc<Self>,
method: Method,
params: EthCallFirstParams,
2023-05-24 00:40:34 +03:00
) -> Web3ProxyResult<()> {
2023-01-19 03:17:43 +03:00
let rpc_key_id = match self.checks.rpc_secret_key_id {
Some(rpc_key_id) => rpc_key_id.into(),
None => {
2023-04-06 01:34:28 +03:00
// trace!(?self, "cannot save revert without rpc_key_id");
return Ok(());
}
};
let db_conn = self.db_conn.as_ref().context("no database connection")?;
// TODO: should the database set the timestamp?
// we intentionally use "now" and not the time the request started
// why? because we aggregate stats and setting one in the past could cause confusion
let timestamp = Utc::now();
let to: Vec<u8> = params
.to
.as_bytes()
.try_into()
.expect("address should always convert to a Vec<u8>");
let call_data = params.data.map(|x| format!("{}", x));
let rl = revert_log::ActiveModel {
rpc_key_id: sea_orm::Set(rpc_key_id),
method: sea_orm::Set(method),
to: sea_orm::Set(to),
call_data: sea_orm::Set(call_data),
timestamp: sea_orm::Set(timestamp),
..Default::default()
};
let rl = rl
.save(db_conn)
.await
2023-06-25 01:24:46 +03:00
.web3_context("Failed saving new revert log")?;
2023-06-25 01:24:46 +03:00
// TODO: what log level and format?
trace!(revert_log=?rl);
// TODO: return something useful
Ok(())
2022-09-23 00:51:52 +03:00
}
}
impl Drop for OpenRequestHandle {
fn drop(&mut self) {
2023-05-13 09:04:56 +03:00
self.rpc
.active_requests
.fetch_sub(1, atomic::Ordering::AcqRel);
}
}
2022-08-24 03:14:49 +03:00
impl OpenRequestHandle {
pub async fn new(
authorization: Arc<Authorization>,
rpc: Arc<Web3Rpc>,
error_handler: Option<RequestErrorHandler>,
) -> Self {
// TODO: take request_id as an argument?
// TODO: attach a unique id to this? customer requests have one, but not internal queries
// TODO: what ordering?!
rpc.active_requests
.fetch_add(1, std::sync::atomic::Ordering::AcqRel);
let error_handler = error_handler.unwrap_or_default();
Self {
authorization,
error_handler,
rpc,
}
2022-08-24 03:11:49 +03:00
}
2022-12-20 02:59:01 +03:00
pub fn connection_name(&self) -> String {
2023-02-12 21:22:20 +03:00
self.rpc.name.clone()
2022-12-20 02:59:01 +03:00
}
2022-09-23 01:34:43 +03:00
#[inline]
pub fn clone_connection(&self) -> Arc<Web3Rpc> {
2023-02-12 21:22:20 +03:00
self.rpc.clone()
2022-08-24 03:11:49 +03:00
}
/// Send a web3 request
/// By having the request method here, we ensure that the rate limiter was called and connection counts were properly incremented
2023-02-16 11:26:58 +03:00
/// depending on how things are locked, you might need to pass the provider in
/// we take self to ensure this function only runs once
pub async fn request<P: JsonRpcParams, R: JsonRpcResultData + serde::Serialize>(
self,
2022-08-24 03:11:49 +03:00
method: &str,
params: &P,
) -> Result<R, ProviderError> {
2022-09-23 00:51:52 +03:00
// TODO: use tracing spans
// TODO: including params in this log is way too verbose
// trace!(rpc=%self.rpc, %method, "request");
2023-02-12 21:22:20 +03:00
trace!("requesting from {}", self.rpc);
2023-06-13 21:51:19 +03:00
match self.authorization.authorization_type {
AuthorizationType::Frontend => {
self.rpc
.external_requests
.fetch_add(1, std::sync::atomic::Ordering::Relaxed);
}
AuthorizationType::Internal => {
self.rpc
.internal_requests
.fetch_add(1, std::sync::atomic::Ordering::Relaxed);
}
}
2023-02-15 23:33:43 +03:00
// we used to fetch_add the active_request count here, but sometimes a request is made without going through this function (like with subscriptions)
2023-02-16 11:26:58 +03:00
let start = Instant::now();
2023-02-15 23:33:43 +03:00
// TODO: replace ethers-rs providers with our own that supports streaming the responses
// TODO: replace ethers-rs providers with our own that handles "id" being null
let response: Result<R, _> = if let Some(ref p) = self.rpc.http_provider {
p.request(method, params).await
} else if let Some(p) = self.rpc.ws_provider.load().as_ref() {
p.request(method, params).await
} else {
2023-05-24 00:51:34 +03:00
return Err(ProviderError::CustomError(
"no provider configured!".to_string(),
));
2022-08-24 03:11:49 +03:00
};
// we do NOT want to measure errors, so we intentionally do not record this latency now.
let latency = start.elapsed();
2023-02-16 11:26:58 +03:00
// we used to fetch_sub the active_request count here, but sometimes the handle is dropped without request being called!
2023-02-16 11:26:58 +03:00
2023-04-24 21:21:53 +03:00
trace!(
"response from {} for {} {:?}: {:?}",
self.rpc,
method,
params,
response,
);
2022-12-06 00:13:36 +03:00
if let Err(err) = &response {
2022-09-23 01:42:44 +03:00
// only save reverts for some types of calls
// TODO: do something special for eth_sendRawTransaction too
let error_handler = if let RequestErrorHandler::Save = self.error_handler {
2022-11-26 07:57:25 +03:00
// TODO: should all these be Trace or Debug or a mix?
2022-09-24 10:04:11 +03:00
if !["eth_call", "eth_estimateGas"].contains(&method) {
2022-11-12 11:24:32 +03:00
// trace!(%method, "skipping save on revert");
2023-04-06 01:34:28 +03:00
RequestErrorHandler::TraceLevel
} else if self.authorization.db_conn.is_some() {
let log_revert_chance = self.authorization.checks.log_revert_chance;
2023-06-13 05:13:06 +03:00
if log_revert_chance == 0 {
2022-11-12 11:24:32 +03:00
// trace!(%method, "no chance. skipping save on revert");
2023-04-06 01:34:28 +03:00
RequestErrorHandler::TraceLevel
2023-06-13 05:13:06 +03:00
} else if log_revert_chance == u16::MAX {
2022-11-12 11:24:32 +03:00
// trace!(%method, "gaurenteed chance. SAVING on revert");
self.error_handler
2023-06-13 21:51:19 +03:00
} else if nanorand::tls_rng().generate_range(0u16..u16::MAX) < log_revert_chance
{
2022-11-12 11:24:32 +03:00
// trace!(%method, "missed chance. skipping save on revert");
2023-04-06 01:34:28 +03:00
RequestErrorHandler::TraceLevel
2022-09-24 10:04:11 +03:00
} else {
2022-11-12 11:24:32 +03:00
// trace!("Saving on revert");
// TODO: is always logging at debug level fine?
self.error_handler
2022-09-24 10:04:11 +03:00
}
2022-09-23 01:42:44 +03:00
} else {
2022-11-12 11:24:32 +03:00
// trace!(%method, "no database. skipping save on revert");
2023-04-06 01:34:28 +03:00
RequestErrorHandler::TraceLevel
}
2022-09-23 01:42:44 +03:00
} else {
self.error_handler
2022-09-23 01:42:44 +03:00
};
// TODO: simple enum -> string derive?
2023-04-06 01:34:28 +03:00
// TODO: if ProviderError::UnsupportedRpc, we should retry on another server
#[derive(Debug)]
2023-04-04 15:40:22 +03:00
enum ResponseTypes {
Revert,
RateLimit,
2023-04-06 01:34:28 +03:00
Error,
}
2022-10-12 00:31:34 +03:00
// check for "execution reverted" here
// TODO: move this info a function on ResponseErrorType
let response_type = if let ProviderError::JsonRpcClientError(err) = err {
2023-06-27 07:11:43 +03:00
// JsonRpc and Application errors get rolled into the JsonRpcClientError
let msg = err.as_error_response().map(|x| x.message.clone());
2022-10-12 00:31:34 +03:00
if let Some(msg) = msg {
2023-06-27 07:11:43 +03:00
trace!(%msg, "jsonrpc error message");
if msg.starts_with("execution reverted") {
2023-02-12 21:22:20 +03:00
trace!("revert from {}", self.rpc);
2023-04-04 15:40:22 +03:00
ResponseTypes::Revert
} else if msg.contains("limit") || msg.contains("request") {
// TODO: too verbose
if self.rpc.backup {
trace!("rate limit from {}", self.rpc);
} else {
warn!("rate limit from {}", self.rpc);
}
2023-04-04 15:40:22 +03:00
ResponseTypes::RateLimit
} else {
2023-04-06 01:34:28 +03:00
ResponseTypes::Error
}
2022-10-12 00:31:34 +03:00
} else {
2023-04-06 01:34:28 +03:00
ResponseTypes::Error
2022-10-12 00:31:34 +03:00
}
} else {
2023-04-06 01:34:28 +03:00
ResponseTypes::Error
2022-10-12 00:31:34 +03:00
};
2022-10-10 07:15:07 +03:00
2023-04-04 15:40:22 +03:00
if matches!(response_type, ResponseTypes::RateLimit) {
if let Some(hard_limit_until) = self.rpc.hard_limit_until.as_ref() {
2023-04-06 01:34:28 +03:00
// TODO: how long should we actually wait? different providers have different times
// TODO: if rate_limit_period_seconds is set, use that
// TODO: check response headers for rate limits too
// TODO: warn if production, debug if backup
if self.rpc.backup {
debug!("unexpected rate limit on {}!", self.rpc);
} else {
warn!("unexpected rate limit on {}!", self.rpc);
}
2023-04-04 15:40:22 +03:00
let retry_at = Instant::now() + Duration::from_secs(1);
2023-04-04 15:40:22 +03:00
trace!("retry {} at: {:?}", self.rpc, retry_at);
2023-04-04 15:40:22 +03:00
hard_limit_until.send_replace(retry_at);
}
2023-04-04 15:40:22 +03:00
}
// TODO: think more about the method and param logs. those can be sensitive information
// we do **NOT** use self.error_handler here because it might have been modified
2023-04-06 01:34:28 +03:00
match error_handler {
RequestErrorHandler::DebugLevel => {
2023-04-04 15:40:22 +03:00
// TODO: think about this revert check more. sometimes we might want reverts logged so this needs a flag
if matches!(response_type, ResponseTypes::Revert) {
2023-06-27 07:11:43 +03:00
trace!(
rpc=%self.rpc,
%method,
?params,
?err,
"revert",
);
} else {
2023-04-04 15:40:22 +03:00
debug!(
2023-06-27 07:11:43 +03:00
rpc=%self.rpc,
%method,
?params,
?err,
"bad response",
2023-04-04 15:40:22 +03:00
);
}
}
2023-04-06 01:34:28 +03:00
RequestErrorHandler::TraceLevel => {
2023-04-04 15:40:22 +03:00
trace!(
2023-06-27 07:11:43 +03:00
rpc=%self.rpc,
%method,
?params,
?err,
"bad response",
2023-04-04 15:40:22 +03:00
);
}
2023-04-06 01:34:28 +03:00
RequestErrorHandler::ErrorLevel => {
2023-06-27 07:11:43 +03:00
// TODO: only include params if not running in release mode
2023-04-04 15:40:22 +03:00
error!(
2023-06-27 07:11:43 +03:00
rpc=%self.rpc,
%method,
?params,
?err,
"bad response",
2023-04-04 15:40:22 +03:00
);
}
2023-04-06 01:34:28 +03:00
RequestErrorHandler::WarnLevel => {
2023-06-27 07:11:43 +03:00
// TODO: only include params if not running in release mode
2023-04-04 15:40:22 +03:00
warn!(
2023-06-27 07:11:43 +03:00
rpc=%self.rpc,
%method,
?params,
?err,
"bad response",
2023-04-04 15:40:22 +03:00
);
}
2023-04-06 01:34:28 +03:00
RequestErrorHandler::Save => {
2023-04-04 15:40:22 +03:00
trace!(
2023-06-27 07:11:43 +03:00
rpc=%self.rpc,
%method,
?params,
?err,
"bad response",
2023-04-04 15:40:22 +03:00
);
// TODO: do not unwrap! (doesn't matter much since we check method as a string above)
let method: Method = Method::try_from_value(&method.to_string()).unwrap();
match serde_json::from_value::<EthCallParams>(json!(params)) {
Ok(params) => {
// spawn saving to the database so we don't slow down the request
let f = self.authorization.clone().save_revert(method, params.0 .0);
2023-04-04 15:40:22 +03:00
tokio::spawn(f);
}
Err(err) => {
2023-05-13 09:04:56 +03:00
warn!(
2023-06-29 04:36:17 +03:00
%method,
?params,
?err,
"failed parsing eth_call params. unable to save revert",
2023-05-13 09:04:56 +03:00
);
}
}
}
}
}
2023-06-20 03:46:58 +03:00
tokio::spawn(async move {
self.rpc.peak_latency.as_ref().unwrap().report(latency);
self.rpc.median_latency.as_ref().unwrap().record(latency);
});
2023-06-18 19:47:40 +03:00
2022-08-24 03:11:49 +03:00
response
}
}