persistence in database
only maybe a little bit shitty but it works very well still needs chunk unloading
This commit is contained in:
parent
2594afcc1b
commit
70e979057c
|
@ -1,4 +1,7 @@
|
||||||
use std::{collections::HashSet, sync::Arc};
|
use std::{
|
||||||
|
collections::{HashSet, VecDeque},
|
||||||
|
sync::Arc,
|
||||||
|
};
|
||||||
|
|
||||||
use axum::{
|
use axum::{
|
||||||
extract::{
|
extract::{
|
||||||
|
@ -13,16 +16,20 @@ use schema::{
|
||||||
ChunkInfo, Error, LoginRequest, LoginResponse, Notify, Online, Request, Version, WallInfo,
|
ChunkInfo, Error, LoginRequest, LoginResponse, Notify, Online, Request, Version, WallInfo,
|
||||||
};
|
};
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use tokio::{select, sync::mpsc, time::Instant};
|
use tokio::{
|
||||||
use tracing::{error, info};
|
select,
|
||||||
|
sync::{self, mpsc, oneshot},
|
||||||
|
time::Instant,
|
||||||
|
};
|
||||||
|
use tracing::{error, info, instrument};
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
haku::{Haku, Limits},
|
haku::{Haku, Limits},
|
||||||
login::database::LoginStatus,
|
login::database::LoginStatus,
|
||||||
schema::Vec2,
|
schema::Vec2,
|
||||||
wall::{
|
wall::{
|
||||||
self, chunk_encoder::ChunkEncoder, chunk_iterator::ChunkIterator, ChunkPosition, JoinError,
|
self, auto_save::AutoSave, chunk_images::ChunkImages, chunk_iterator::ChunkIterator,
|
||||||
SessionHandle, UserInit, Wall,
|
database::ChunkDataPair, ChunkPosition, JoinError, SessionHandle, UserInit, Wall,
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -110,7 +117,7 @@ async fn fallible_websocket(api: Arc<Api>, ws: &mut WebSocket) -> eyre::Result<(
|
||||||
Some(wall) => wall,
|
Some(wall) => wall,
|
||||||
None => api.dbs.wall_broker.generate_id().await,
|
None => api.dbs.wall_broker.generate_id().await,
|
||||||
};
|
};
|
||||||
let open_wall = api.dbs.wall_broker.open(wall_id);
|
let open_wall = api.dbs.wall_broker.open(wall_id).await?;
|
||||||
|
|
||||||
let session_handle = match open_wall
|
let session_handle = match open_wall
|
||||||
.wall
|
.wall
|
||||||
|
@ -178,7 +185,8 @@ async fn fallible_websocket(api: Arc<Api>, ws: &mut WebSocket) -> eyre::Result<(
|
||||||
|
|
||||||
SessionLoop::start(
|
SessionLoop::start(
|
||||||
open_wall.wall,
|
open_wall.wall,
|
||||||
open_wall.chunk_encoder,
|
open_wall.chunk_images,
|
||||||
|
open_wall.auto_save,
|
||||||
session_handle,
|
session_handle,
|
||||||
api.config.haku.clone(),
|
api.config.haku.clone(),
|
||||||
login_request.init.brush,
|
login_request.init.brush,
|
||||||
|
@ -192,24 +200,33 @@ async fn fallible_websocket(api: Arc<Api>, ws: &mut WebSocket) -> eyre::Result<(
|
||||||
|
|
||||||
struct SessionLoop {
|
struct SessionLoop {
|
||||||
wall: Arc<Wall>,
|
wall: Arc<Wall>,
|
||||||
chunk_encoder: Arc<ChunkEncoder>,
|
chunk_images: Arc<ChunkImages>,
|
||||||
|
auto_save: Arc<AutoSave>,
|
||||||
handle: SessionHandle,
|
handle: SessionHandle,
|
||||||
|
|
||||||
render_commands_tx: mpsc::Sender<RenderCommand>,
|
render_commands_tx: mpsc::Sender<RenderCommand>,
|
||||||
|
|
||||||
viewport_chunks: ChunkIterator,
|
viewport_chunks: ChunkIterator,
|
||||||
sent_chunks: HashSet<ChunkPosition>,
|
sent_chunks: HashSet<ChunkPosition>,
|
||||||
|
pending_images: VecDeque<ChunkDataPair>,
|
||||||
}
|
}
|
||||||
|
|
||||||
enum RenderCommand {
|
enum RenderCommand {
|
||||||
SetBrush { brush: String },
|
SetBrush {
|
||||||
Plot { points: Vec<Vec2> },
|
brush: String,
|
||||||
|
},
|
||||||
|
|
||||||
|
Plot {
|
||||||
|
points: Vec<Vec2>,
|
||||||
|
done: oneshot::Sender<()>,
|
||||||
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
impl SessionLoop {
|
impl SessionLoop {
|
||||||
async fn start(
|
async fn start(
|
||||||
wall: Arc<Wall>,
|
wall: Arc<Wall>,
|
||||||
chunk_encoder: Arc<ChunkEncoder>,
|
chunk_images: Arc<ChunkImages>,
|
||||||
|
auto_save: Arc<AutoSave>,
|
||||||
handle: SessionHandle,
|
handle: SessionHandle,
|
||||||
limits: Limits,
|
limits: Limits,
|
||||||
brush: String,
|
brush: String,
|
||||||
|
@ -230,18 +247,20 @@ impl SessionLoop {
|
||||||
.name(String::from("haku render thread"))
|
.name(String::from("haku render thread"))
|
||||||
.spawn({
|
.spawn({
|
||||||
let wall = Arc::clone(&wall);
|
let wall = Arc::clone(&wall);
|
||||||
let chunk_encoder = Arc::clone(&chunk_encoder);
|
let chunk_images = Arc::clone(&chunk_images);
|
||||||
move || Self::render_thread(wall, chunk_encoder, limits, render_commands_rx)
|
move || Self::render_thread(wall, chunk_images, limits, render_commands_rx)
|
||||||
})
|
})
|
||||||
.context("could not spawn render thread")?;
|
.context("could not spawn render thread")?;
|
||||||
|
|
||||||
Ok(Self {
|
Ok(Self {
|
||||||
wall,
|
wall,
|
||||||
chunk_encoder,
|
chunk_images,
|
||||||
|
auto_save,
|
||||||
handle,
|
handle,
|
||||||
render_commands_tx,
|
render_commands_tx,
|
||||||
viewport_chunks: ChunkIterator::new(ChunkPosition::new(0, 0), ChunkPosition::new(0, 0)),
|
viewport_chunks: ChunkIterator::new(ChunkPosition::new(0, 0), ChunkPosition::new(0, 0)),
|
||||||
sent_chunks: HashSet::new(),
|
sent_chunks: HashSet::new(),
|
||||||
|
pending_images: VecDeque::new(),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -284,13 +303,28 @@ impl SessionLoop {
|
||||||
.await;
|
.await;
|
||||||
}
|
}
|
||||||
wall::EventKind::Plot { points } => {
|
wall::EventKind::Plot { points } => {
|
||||||
|
let chunks_to_modify: Vec<_> =
|
||||||
|
chunks_to_modify(&self.wall, points).into_iter().collect();
|
||||||
|
match self.chunk_images.load(chunks_to_modify.clone()).await {
|
||||||
|
Ok(_) => {
|
||||||
// We drop commands if we take too long to render instead of lagging
|
// We drop commands if we take too long to render instead of lagging
|
||||||
// the WebSocket thread.
|
// the WebSocket thread.
|
||||||
// Theoretically this will yield much better responsiveness, but it _will_
|
// Theoretically this will yield much better responsiveness, but it _will_
|
||||||
// result in some visual glitches if we're getting bottlenecked.
|
// result in some visual glitches if we're getting bottlenecked.
|
||||||
|
let (done_tx, done_rx) = oneshot::channel();
|
||||||
_ = self.render_commands_tx.try_send(RenderCommand::Plot {
|
_ = self.render_commands_tx.try_send(RenderCommand::Plot {
|
||||||
points: points.clone(),
|
points: points.clone(),
|
||||||
})
|
done: done_tx,
|
||||||
|
});
|
||||||
|
|
||||||
|
let auto_save = Arc::clone(&self.auto_save);
|
||||||
|
tokio::spawn(async move {
|
||||||
|
_ = done_rx.await;
|
||||||
|
auto_save.request(chunks_to_modify).await;
|
||||||
|
});
|
||||||
|
}
|
||||||
|
Err(err) => error!(?err, "while loading chunks for render command"),
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -320,47 +354,53 @@ impl SessionLoop {
|
||||||
let mut chunk_infos = vec![];
|
let mut chunk_infos = vec![];
|
||||||
let mut packet = vec![];
|
let mut packet = vec![];
|
||||||
|
|
||||||
|
if self.pending_images.is_empty() {
|
||||||
|
let mut positions = vec![];
|
||||||
|
|
||||||
// Number of chunks iterated is limited per packet, so as not to let the client
|
// Number of chunks iterated is limited per packet, so as not to let the client
|
||||||
// stall the server by sending in a huge viewport.
|
// stall the server by sending in a huge viewport.
|
||||||
let start = Instant::now();
|
for _ in 0..9000 {
|
||||||
let mut iterated = 0;
|
|
||||||
for i in 0..12000 {
|
|
||||||
if let Some(position) = self.viewport_chunks.next() {
|
if let Some(position) = self.viewport_chunks.next() {
|
||||||
if self.sent_chunks.contains(&position) {
|
if !self.sent_chunks.insert(position)
|
||||||
|
|| !self.chunk_images.chunk_exists(position)
|
||||||
|
{
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
positions.push(position);
|
||||||
|
} else {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
if let Some(encoded) = self.chunk_encoder.encoded(position).await {
|
self.pending_images
|
||||||
|
.extend(self.chunk_images.encoded(positions).await);
|
||||||
|
}
|
||||||
|
|
||||||
|
while let Some(ChunkDataPair { position, data }) = self.pending_images.pop_front() {
|
||||||
let offset = packet.len();
|
let offset = packet.len();
|
||||||
packet.extend_from_slice(&encoded);
|
packet.extend_from_slice(&data);
|
||||||
chunk_infos.push(ChunkInfo {
|
chunk_infos.push(ChunkInfo {
|
||||||
position,
|
position,
|
||||||
offset: u32::try_from(offset).context("packet too big")?,
|
offset: u32::try_from(offset).context("packet too big")?,
|
||||||
length: u32::try_from(encoded.len()).context("chunk image too big")?,
|
length: u32::try_from(data.len()).context("chunk image too big")?,
|
||||||
});
|
});
|
||||||
|
|
||||||
// The actual number of chunks per packet is limited by the packet's size, which
|
// The final number of chunks per packet is limited by the packet's size, which
|
||||||
// we don't want to be too big, to maintain responsiveness - the client will
|
// we don't want to be too big, to maintain responsiveness - the client will
|
||||||
// only request more chunks once per frame, so interactions still have time to
|
// only request more chunks once per frame, so interactions still have time to
|
||||||
// execute. We cap it to 256KiB in hopes that noone has Internet slow enough for
|
// execute. We cap it to 256KiB in hopes that noone has Internet slow enough for
|
||||||
// this to cause a disconnect.
|
// this to cause a disconnect.
|
||||||
|
//
|
||||||
|
// Note that after this there _may_ be more chunks pending in the queue.
|
||||||
if packet.len() >= 256 * 1024 {
|
if packet.len() >= 256 * 1024 {
|
||||||
iterated = i;
|
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
self.sent_chunks.insert(position);
|
|
||||||
} else {
|
|
||||||
iterated = i;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
info!(elapsed = ?start.elapsed(), iterated, "send_chunks");
|
|
||||||
|
|
||||||
ws.send(to_message(&Notify::Chunks {
|
ws.send(to_message(&Notify::Chunks {
|
||||||
chunks: chunk_infos,
|
chunks: chunk_infos,
|
||||||
has_more: self.viewport_chunks.clone().next().is_some(),
|
has_more: !self.pending_images.is_empty()
|
||||||
|
|| self.viewport_chunks.clone().next().is_some(),
|
||||||
}))
|
}))
|
||||||
.await?;
|
.await?;
|
||||||
ws.send(Message::Binary(packet)).await?;
|
ws.send(Message::Binary(packet)).await?;
|
||||||
|
@ -370,7 +410,7 @@ impl SessionLoop {
|
||||||
|
|
||||||
fn render_thread(
|
fn render_thread(
|
||||||
wall: Arc<Wall>,
|
wall: Arc<Wall>,
|
||||||
chunk_encoder: Arc<ChunkEncoder>,
|
chunk_images: Arc<ChunkImages>,
|
||||||
limits: Limits,
|
limits: Limits,
|
||||||
mut commands: mpsc::Receiver<RenderCommand>,
|
mut commands: mpsc::Receiver<RenderCommand>,
|
||||||
) {
|
) {
|
||||||
|
@ -382,29 +422,51 @@ impl SessionLoop {
|
||||||
RenderCommand::SetBrush { brush } => {
|
RenderCommand::SetBrush { brush } => {
|
||||||
brush_ok = haku.set_brush(&brush).is_ok();
|
brush_ok = haku.set_brush(&brush).is_ok();
|
||||||
}
|
}
|
||||||
RenderCommand::Plot { points } => {
|
|
||||||
|
RenderCommand::Plot { points, done } => {
|
||||||
if brush_ok {
|
if brush_ok {
|
||||||
if let Ok(value) = haku.eval_brush() {
|
if let Ok(value) = haku.eval_brush() {
|
||||||
for point in points {
|
for point in points {
|
||||||
// Ignore the result. It's better if we render _something_ rather
|
// Ignore the result. It's better if we render _something_ rather
|
||||||
// than nothing.
|
// than nothing.
|
||||||
_ = draw_to_chunks(&haku, value, point, &wall, &chunk_encoder);
|
_ = draw_to_chunks(&wall, &chunk_images, &haku, value, point);
|
||||||
}
|
}
|
||||||
haku.reset_vm();
|
haku.reset_vm();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
_ = done.send(());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn chunks_to_modify(wall: &Wall, points: &[Vec2]) -> HashSet<ChunkPosition> {
|
||||||
|
let mut chunks = HashSet::new();
|
||||||
|
for point in points {
|
||||||
|
let paint_area = wall.settings().paint_area as f32;
|
||||||
|
let left = point.x - paint_area / 2.0;
|
||||||
|
let top = point.y - paint_area / 2.0;
|
||||||
|
let top_left_chunk = wall.settings().chunk_at(Vec2::new(left, top));
|
||||||
|
let bottom_right_chunk = wall
|
||||||
|
.settings()
|
||||||
|
.chunk_at(Vec2::new(left + paint_area, top + paint_area));
|
||||||
|
for chunk_y in top_left_chunk.y..bottom_right_chunk.y {
|
||||||
|
for chunk_x in top_left_chunk.x..bottom_right_chunk.x {
|
||||||
|
chunks.insert(ChunkPosition::new(chunk_x, chunk_y));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
chunks
|
||||||
|
}
|
||||||
|
|
||||||
|
#[instrument(skip(wall, chunk_images, haku, value))]
|
||||||
fn draw_to_chunks(
|
fn draw_to_chunks(
|
||||||
|
wall: &Wall,
|
||||||
|
chunk_images: &ChunkImages,
|
||||||
haku: &Haku,
|
haku: &Haku,
|
||||||
value: Value,
|
value: Value,
|
||||||
center: Vec2,
|
center: Vec2,
|
||||||
wall: &Wall,
|
|
||||||
chunk_encoder: &ChunkEncoder,
|
|
||||||
) -> eyre::Result<()> {
|
) -> eyre::Result<()> {
|
||||||
let settings = wall.settings();
|
let settings = wall.settings();
|
||||||
|
|
||||||
|
@ -414,10 +476,10 @@ fn draw_to_chunks(
|
||||||
let left = center.x - paint_area / 2.0;
|
let left = center.x - paint_area / 2.0;
|
||||||
let top = center.y - paint_area / 2.0;
|
let top = center.y - paint_area / 2.0;
|
||||||
|
|
||||||
let left_chunk = f32::floor(left / chunk_size) as i32;
|
let left_chunk = settings.chunk_at_1d(left);
|
||||||
let top_chunk = f32::floor(top / chunk_size) as i32;
|
let top_chunk = settings.chunk_at_1d(top);
|
||||||
let right_chunk = f32::ceil((left + paint_area) / chunk_size) as i32;
|
let right_chunk = settings.chunk_at_1d(left + paint_area);
|
||||||
let bottom_chunk = f32::ceil((top + paint_area) / chunk_size) as i32;
|
let bottom_chunk = settings.chunk_at_1d(top + paint_area);
|
||||||
for chunk_y in top_chunk..bottom_chunk {
|
for chunk_y in top_chunk..bottom_chunk {
|
||||||
for chunk_x in left_chunk..right_chunk {
|
for chunk_x in left_chunk..right_chunk {
|
||||||
let x = f32::floor(-chunk_x as f32 * chunk_size + center.x);
|
let x = f32::floor(-chunk_x as f32 * chunk_size + center.x);
|
||||||
|
@ -428,11 +490,16 @@ fn draw_to_chunks(
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
for chunk_y in top_chunk..bottom_chunk {
|
// NOTE: Maybe sending in an iterator would be more efficient?
|
||||||
for chunk_x in left_chunk..right_chunk {
|
// If there were many chunks modified, (which there probably weren't,) this could allocate
|
||||||
chunk_encoder.invalidate_blocking(ChunkPosition::new(chunk_x, chunk_y))
|
// a lot of memory.
|
||||||
}
|
chunk_images.mark_modified_blocking(
|
||||||
}
|
(top_chunk..bottom_chunk)
|
||||||
|
.flat_map(|chunk_y| {
|
||||||
|
(left_chunk..right_chunk).map(move |chunk_x| ChunkPosition::new(chunk_x, chunk_y))
|
||||||
|
})
|
||||||
|
.collect(),
|
||||||
|
);
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
|
@ -4,6 +4,6 @@ use crate::wall;
|
||||||
|
|
||||||
#[derive(Debug, Clone, Deserialize, Serialize)]
|
#[derive(Debug, Clone, Deserialize, Serialize)]
|
||||||
pub struct Config {
|
pub struct Config {
|
||||||
pub wall: wall::Settings,
|
pub wall_broker: wall::broker::Settings,
|
||||||
pub haku: crate::haku::Limits,
|
pub haku: crate::haku::Limits,
|
||||||
}
|
}
|
||||||
|
|
|
@ -14,7 +14,7 @@ use serde::{Deserialize, Serialize};
|
||||||
use crate::{id, serialization::DeserializeFromStr};
|
use crate::{id, serialization::DeserializeFromStr};
|
||||||
|
|
||||||
#[derive(Debug, Clone, Copy, PartialEq, Eq)]
|
#[derive(Debug, Clone, Copy, PartialEq, Eq)]
|
||||||
pub struct UserId([u8; 32]);
|
pub struct UserId(pub [u8; 32]);
|
||||||
|
|
||||||
impl UserId {
|
impl UserId {
|
||||||
pub fn new(rng: &mut dyn RngCore) -> Self {
|
pub fn new(rng: &mut dyn RngCore) -> Self {
|
||||||
|
|
|
@ -18,6 +18,7 @@ pub struct Database {
|
||||||
command_tx: mpsc::Sender<Command>,
|
command_tx: mpsc::Sender<Command>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Copy, PartialEq, Eq)]
|
||||||
pub enum LoginStatus {
|
pub enum LoginStatus {
|
||||||
ValidUser,
|
ValidUser,
|
||||||
UserDoesNotExist,
|
UserDoesNotExist,
|
||||||
|
@ -33,10 +34,12 @@ enum Command {
|
||||||
nickname: String,
|
nickname: String,
|
||||||
reply: oneshot::Sender<eyre::Result<UserId>>,
|
reply: oneshot::Sender<eyre::Result<UserId>>,
|
||||||
},
|
},
|
||||||
|
|
||||||
LogIn {
|
LogIn {
|
||||||
user_id: UserId,
|
user_id: UserId,
|
||||||
reply: oneshot::Sender<LoginStatus>,
|
reply: oneshot::Sender<LoginStatus>,
|
||||||
},
|
},
|
||||||
|
|
||||||
UserInfo {
|
UserInfo {
|
||||||
user_id: UserId,
|
user_id: UserId,
|
||||||
reply: oneshot::Sender<eyre::Result<Option<UserInfo>>>,
|
reply: oneshot::Sender<eyre::Result<Option<UserInfo>>>,
|
||||||
|
@ -79,8 +82,10 @@ impl Database {
|
||||||
pub fn start(settings: &Settings) -> eyre::Result<Database> {
|
pub fn start(settings: &Settings) -> eyre::Result<Database> {
|
||||||
let db = Connection::open(&settings.path).context("cannot open login database")?;
|
let db = Connection::open(&settings.path).context("cannot open login database")?;
|
||||||
|
|
||||||
db.execute(
|
db.execute_batch(
|
||||||
r#"
|
r#"
|
||||||
|
PRAGMA application_id = 0x726B674C; -- rkgL
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS
|
CREATE TABLE IF NOT EXISTS
|
||||||
t_users (
|
t_users (
|
||||||
user_index INTEGER PRIMARY KEY,
|
user_index INTEGER PRIMARY KEY,
|
||||||
|
@ -89,14 +94,15 @@ pub fn start(settings: &Settings) -> eyre::Result<Database> {
|
||||||
last_login_timestamp INTEGER NOT NULL
|
last_login_timestamp INTEGER NOT NULL
|
||||||
);
|
);
|
||||||
"#,
|
"#,
|
||||||
(),
|
|
||||||
)?;
|
)?;
|
||||||
|
|
||||||
let (command_tx, mut command_rx) = mpsc::channel(8);
|
let (command_tx, mut command_rx) = mpsc::channel(8);
|
||||||
|
|
||||||
let mut user_id_rng = rand_chacha::ChaCha20Rng::from_entropy();
|
let mut user_id_rng = rand_chacha::ChaCha20Rng::from_entropy();
|
||||||
|
|
||||||
tokio::task::spawn_blocking(move || {
|
std::thread::Builder::new()
|
||||||
|
.name("login database thread".into())
|
||||||
|
.spawn(move || {
|
||||||
let mut s_insert_user = db
|
let mut s_insert_user = db
|
||||||
.prepare(
|
.prepare(
|
||||||
r#"
|
r#"
|
||||||
|
@ -140,7 +146,8 @@ pub fn start(settings: &Settings) -> eyre::Result<Database> {
|
||||||
|
|
||||||
Command::LogIn { user_id, reply } => {
|
Command::LogIn { user_id, reply } => {
|
||||||
// TODO: User expiration.
|
// TODO: User expiration.
|
||||||
let login_status = match s_log_in.execute((Utc::now().timestamp(), user_id.0)) {
|
let login_status =
|
||||||
|
match s_log_in.execute((Utc::now().timestamp(), user_id.0)) {
|
||||||
Ok(_) => LoginStatus::ValidUser,
|
Ok(_) => LoginStatus::ValidUser,
|
||||||
Err(_) => LoginStatus::UserDoesNotExist,
|
Err(_) => LoginStatus::UserDoesNotExist,
|
||||||
};
|
};
|
||||||
|
@ -160,7 +167,8 @@ pub fn start(settings: &Settings) -> eyre::Result<Database> {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
});
|
})
|
||||||
|
.context("cannot spawn thread")?;
|
||||||
|
|
||||||
Ok(Database { command_tx })
|
Ok(Database { command_tx })
|
||||||
}
|
}
|
||||||
|
|
|
@ -67,7 +67,9 @@ fn database(config: &Config, paths: &Paths<'_>) -> eyre::Result<Databases> {
|
||||||
})
|
})
|
||||||
.context("cannot start up login database")?;
|
.context("cannot start up login database")?;
|
||||||
|
|
||||||
let wall_broker = wall::Broker::new(config.wall);
|
create_dir_all(paths.database_dir.join("wall"))?;
|
||||||
|
let wall_broker =
|
||||||
|
wall::Broker::new(paths.database_dir.join("wall"), config.wall_broker.clone());
|
||||||
|
|
||||||
Ok(Databases { login, wall_broker })
|
Ok(Databases { login, wall_broker })
|
||||||
}
|
}
|
||||||
|
@ -89,7 +91,6 @@ async fn fallible_main() -> eyre::Result<()> {
|
||||||
let dbs = Arc::new(database(&config, &paths)?);
|
let dbs = Arc::new(database(&config, &paths)?);
|
||||||
|
|
||||||
let api = Arc::new(Api { config, dbs });
|
let api = Arc::new(Api { config, dbs });
|
||||||
|
|
||||||
let app = Router::new()
|
let app = Router::new()
|
||||||
.route_service(
|
.route_service(
|
||||||
"/",
|
"/",
|
||||||
|
@ -116,9 +117,7 @@ async fn main() {
|
||||||
let _client = tracy_client::Client::start();
|
let _client = tracy_client::Client::start();
|
||||||
|
|
||||||
color_eyre::install().unwrap();
|
color_eyre::install().unwrap();
|
||||||
tracing_subscriber::fmt()
|
tracing_subscriber::fmt().init();
|
||||||
.with_span_events(FmtSpan::ACTIVE)
|
|
||||||
.init();
|
|
||||||
|
|
||||||
match fallible_main().await {
|
match fallible_main().await {
|
||||||
Ok(_) => (),
|
Ok(_) => (),
|
||||||
|
|
|
@ -5,3 +5,9 @@ pub struct Vec2 {
|
||||||
pub x: f32,
|
pub x: f32,
|
||||||
pub y: f32,
|
pub y: f32,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl Vec2 {
|
||||||
|
pub fn new(x: f32, y: f32) -> Self {
|
||||||
|
Self { x, y }
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
|
@ -17,11 +17,14 @@ use tracing::info;
|
||||||
|
|
||||||
use crate::{id, login::UserId, schema::Vec2, serialization::DeserializeFromStr};
|
use crate::{id, login::UserId, schema::Vec2, serialization::DeserializeFromStr};
|
||||||
|
|
||||||
|
pub mod auto_save;
|
||||||
pub mod broker;
|
pub mod broker;
|
||||||
pub mod chunk_encoder;
|
pub mod chunk_images;
|
||||||
pub mod chunk_iterator;
|
pub mod chunk_iterator;
|
||||||
|
pub mod database;
|
||||||
|
|
||||||
pub use broker::Broker;
|
pub use broker::Broker;
|
||||||
|
pub use database::Database;
|
||||||
|
|
||||||
#[derive(Debug, Clone, Copy, PartialEq, Eq, Hash)]
|
#[derive(Debug, Clone, Copy, PartialEq, Eq, Hash)]
|
||||||
pub struct WallId([u8; 32]);
|
pub struct WallId([u8; 32]);
|
||||||
|
@ -82,12 +85,18 @@ impl fmt::Display for InvalidWallId {
|
||||||
|
|
||||||
impl Error for InvalidWallId {}
|
impl Error for InvalidWallId {}
|
||||||
|
|
||||||
#[derive(Debug, Clone, Copy, PartialEq, Eq, Hash, Deserialize, Serialize)]
|
#[derive(Clone, Copy, PartialEq, Eq, Hash, Deserialize, Serialize)]
|
||||||
pub struct ChunkPosition {
|
pub struct ChunkPosition {
|
||||||
pub x: i32,
|
pub x: i32,
|
||||||
pub y: i32,
|
pub y: i32,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl fmt::Debug for ChunkPosition {
|
||||||
|
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||||
|
write!(f, "({}, {})", self.x, self.y)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl ChunkPosition {
|
impl ChunkPosition {
|
||||||
pub fn new(x: i32, y: i32) -> Self {
|
pub fn new(x: i32, y: i32) -> Self {
|
||||||
Self { x, y }
|
Self { x, y }
|
||||||
|
@ -115,11 +124,12 @@ pub struct Settings {
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Settings {
|
impl Settings {
|
||||||
|
pub fn chunk_at_1d(&self, x: f32) -> i32 {
|
||||||
|
f32::floor(x / self.chunk_size as f32) as i32
|
||||||
|
}
|
||||||
|
|
||||||
pub fn chunk_at(&self, position: Vec2) -> ChunkPosition {
|
pub fn chunk_at(&self, position: Vec2) -> ChunkPosition {
|
||||||
ChunkPosition::new(
|
ChunkPosition::new(self.chunk_at_1d(position.x), self.chunk_at_1d(position.y))
|
||||||
f32::floor(position.x / self.chunk_size as f32) as i32,
|
|
||||||
f32::floor(position.y / self.chunk_size as f32) as i32,
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -200,17 +210,19 @@ impl Wall {
|
||||||
&self.settings
|
&self.settings
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn has_chunk(&self, at: ChunkPosition) -> bool {
|
||||||
|
self.chunks.contains_key(&at)
|
||||||
|
}
|
||||||
|
|
||||||
pub fn get_chunk(&self, at: ChunkPosition) -> Option<Arc<Mutex<Chunk>>> {
|
pub fn get_chunk(&self, at: ChunkPosition) -> Option<Arc<Mutex<Chunk>>> {
|
||||||
self.chunks.get(&at).map(|chunk| Arc::clone(&chunk))
|
self.chunks.get(&at).map(|chunk| Arc::clone(&chunk))
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_or_create_chunk(&self, at: ChunkPosition) -> Arc<Mutex<Chunk>> {
|
pub fn get_or_create_chunk(&self, at: ChunkPosition) -> Arc<Mutex<Chunk>> {
|
||||||
Arc::clone(
|
Arc::clone(&self.chunks.entry(at).or_insert_with(|| {
|
||||||
&self
|
info!(?at, "chunk created");
|
||||||
.chunks
|
Arc::new(Mutex::new(Chunk::new(self.settings.chunk_size)))
|
||||||
.entry(at)
|
}))
|
||||||
.or_insert_with(|| Arc::new(Mutex::new(Chunk::new(self.settings.chunk_size)))),
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn join(self: &Arc<Self>, session: Session) -> Result<SessionHandle, JoinError> {
|
pub fn join(self: &Arc<Self>, session: Session) -> Result<SessionHandle, JoinError> {
|
||||||
|
|
86
crates/rkgk/src/wall/auto_save.rs
Normal file
86
crates/rkgk/src/wall/auto_save.rs
Normal file
|
@ -0,0 +1,86 @@
|
||||||
|
use std::{backtrace::Backtrace, collections::HashSet, sync::Arc, time::Duration};
|
||||||
|
|
||||||
|
use dashmap::DashSet;
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
use tokio::{
|
||||||
|
sync::mpsc,
|
||||||
|
time::{interval, MissedTickBehavior},
|
||||||
|
};
|
||||||
|
use tracing::{info, instrument};
|
||||||
|
|
||||||
|
use super::{chunk_images::ChunkImages, ChunkPosition, Database, Wall};
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Deserialize, Serialize)]
|
||||||
|
pub struct Settings {
|
||||||
|
pub interval_seconds: u64,
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct AutoSave {
|
||||||
|
requests_tx: mpsc::Sender<Vec<ChunkPosition>>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl AutoSave {
|
||||||
|
pub fn new(chunk_images: Arc<ChunkImages>, settings: Settings) -> Self {
|
||||||
|
let (requests_tx, requests_rx) = mpsc::channel(8);
|
||||||
|
|
||||||
|
tokio::spawn(
|
||||||
|
AutoSaveLoop {
|
||||||
|
chunk_images,
|
||||||
|
settings,
|
||||||
|
requests_rx,
|
||||||
|
unsaved_chunks: HashSet::new(),
|
||||||
|
}
|
||||||
|
.enter(),
|
||||||
|
);
|
||||||
|
|
||||||
|
Self { requests_tx }
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn request(&self, chunks: Vec<ChunkPosition>) {
|
||||||
|
_ = self.requests_tx.send(chunks).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
struct AutoSaveLoop {
|
||||||
|
chunk_images: Arc<ChunkImages>,
|
||||||
|
settings: Settings,
|
||||||
|
|
||||||
|
requests_rx: mpsc::Receiver<Vec<ChunkPosition>>,
|
||||||
|
unsaved_chunks: HashSet<ChunkPosition>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl AutoSaveLoop {
|
||||||
|
async fn enter(mut self) {
|
||||||
|
let mut save_interval = interval(Duration::from_secs(self.settings.interval_seconds));
|
||||||
|
save_interval.set_missed_tick_behavior(MissedTickBehavior::Skip);
|
||||||
|
|
||||||
|
loop {
|
||||||
|
tokio::select! {
|
||||||
|
request = self.requests_rx.recv() => {
|
||||||
|
if let Some(positions) = request {
|
||||||
|
for position in positions {
|
||||||
|
self.unsaved_chunks.insert(position);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
_ = save_interval.tick() => self.save_chunks().await,
|
||||||
|
|
||||||
|
else => break,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[instrument(skip(self), fields(num_chunks = self.unsaved_chunks.len()))]
|
||||||
|
async fn save_chunks(&mut self) {
|
||||||
|
// NOTE: We don't care about actually using the images here -
|
||||||
|
// the ChunkImages service writes them to the database by itself, and that's all our
|
||||||
|
// request is for.
|
||||||
|
_ = self
|
||||||
|
.chunk_images
|
||||||
|
.encoded(self.unsaved_chunks.drain().collect())
|
||||||
|
.await;
|
||||||
|
}
|
||||||
|
}
|
|
@ -1,19 +1,31 @@
|
||||||
use std::sync::Arc;
|
use std::{path::PathBuf, sync::Arc};
|
||||||
|
|
||||||
use dashmap::DashMap;
|
use dashmap::DashMap;
|
||||||
use rand::SeedableRng;
|
use rand::SeedableRng;
|
||||||
use rand_chacha::ChaCha20Rng;
|
use rand_chacha::ChaCha20Rng;
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
use tokio::sync::Mutex;
|
use tokio::sync::Mutex;
|
||||||
use tracing::info;
|
use tracing::{info, instrument};
|
||||||
|
|
||||||
use super::{chunk_encoder::ChunkEncoder, Settings, Wall, WallId};
|
use super::{
|
||||||
|
auto_save::{self, AutoSave},
|
||||||
|
chunk_images::ChunkImages,
|
||||||
|
database, Database, Wall, WallId,
|
||||||
|
};
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Deserialize, Serialize)]
|
||||||
|
pub struct Settings {
|
||||||
|
pub default_wall_settings: super::Settings,
|
||||||
|
pub auto_save: auto_save::Settings,
|
||||||
|
}
|
||||||
|
|
||||||
/// The broker is the main way to access wall data.
|
/// The broker is the main way to access wall data.
|
||||||
///
|
///
|
||||||
/// It handles dynamically loading and unloading walls as they're needed.
|
/// It handles dynamically loading and unloading walls as they're needed.
|
||||||
/// It also handles database threads for each wall.
|
/// It also handles database threads for each wall.
|
||||||
pub struct Broker {
|
pub struct Broker {
|
||||||
wall_settings: Settings,
|
databases_dir: PathBuf,
|
||||||
|
settings: Settings,
|
||||||
open_walls: DashMap<WallId, OpenWall>,
|
open_walls: DashMap<WallId, OpenWall>,
|
||||||
rng: Mutex<ChaCha20Rng>,
|
rng: Mutex<ChaCha20Rng>,
|
||||||
}
|
}
|
||||||
|
@ -21,14 +33,17 @@ pub struct Broker {
|
||||||
#[derive(Clone)]
|
#[derive(Clone)]
|
||||||
pub struct OpenWall {
|
pub struct OpenWall {
|
||||||
pub wall: Arc<Wall>,
|
pub wall: Arc<Wall>,
|
||||||
pub chunk_encoder: Arc<ChunkEncoder>,
|
pub chunk_images: Arc<ChunkImages>,
|
||||||
|
pub db: Arc<Database>,
|
||||||
|
pub auto_save: Arc<AutoSave>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Broker {
|
impl Broker {
|
||||||
pub fn new(wall_settings: Settings) -> Self {
|
pub fn new(databases_dir: PathBuf, settings: Settings) -> Self {
|
||||||
info!(?wall_settings, "Broker::new");
|
info!(?settings, "Broker::new");
|
||||||
Self {
|
Self {
|
||||||
wall_settings,
|
databases_dir,
|
||||||
|
settings,
|
||||||
open_walls: DashMap::new(),
|
open_walls: DashMap::new(),
|
||||||
rng: Mutex::new(ChaCha20Rng::from_entropy()),
|
rng: Mutex::new(ChaCha20Rng::from_entropy()),
|
||||||
}
|
}
|
||||||
|
@ -41,14 +56,35 @@ impl Broker {
|
||||||
WallId::new(&mut *rng)
|
WallId::new(&mut *rng)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn open(&self, wall_id: WallId) -> OpenWall {
|
#[instrument(skip(self), fields(%wall_id))]
|
||||||
let wall = Arc::new(Wall::new(self.wall_settings));
|
pub async fn open(&self, wall_id: WallId) -> eyre::Result<OpenWall> {
|
||||||
self.open_walls
|
let open_wall = self.open_walls.entry(wall_id);
|
||||||
.entry(wall_id)
|
|
||||||
.or_insert_with(|| OpenWall {
|
match open_wall {
|
||||||
chunk_encoder: Arc::new(ChunkEncoder::start(Arc::clone(&wall))),
|
dashmap::Entry::Vacant(entry) => {
|
||||||
|
let db = Arc::new(database::start(database::Settings {
|
||||||
|
path: self.databases_dir.join(format!("{wall_id}.db")),
|
||||||
|
wall_id,
|
||||||
|
default_wall_settings: self.settings.default_wall_settings,
|
||||||
|
})?);
|
||||||
|
let wall = Arc::new(Wall::new(*db.wall_settings()));
|
||||||
|
let chunk_images = Arc::new(ChunkImages::new(Arc::clone(&wall), Arc::clone(&db)));
|
||||||
|
let auto_save = Arc::new(AutoSave::new(
|
||||||
|
Arc::clone(&chunk_images),
|
||||||
|
self.settings.auto_save.clone(),
|
||||||
|
));
|
||||||
|
let open_wall = OpenWall {
|
||||||
wall,
|
wall,
|
||||||
})
|
chunk_images,
|
||||||
.clone()
|
db,
|
||||||
|
auto_save,
|
||||||
|
};
|
||||||
|
|
||||||
|
entry.insert(open_wall.clone());
|
||||||
|
|
||||||
|
Ok(open_wall)
|
||||||
|
}
|
||||||
|
dashmap::Entry::Occupied(entry) => Ok(entry.get().clone()),
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,104 +0,0 @@
|
||||||
use std::sync::Arc;
|
|
||||||
|
|
||||||
use indexmap::IndexMap;
|
|
||||||
use tokio::sync::{mpsc, oneshot};
|
|
||||||
|
|
||||||
use super::{ChunkPosition, Wall};
|
|
||||||
|
|
||||||
/// Service which encodes chunks to WebP images and caches them in an LRU fashion.
|
|
||||||
pub struct ChunkEncoder {
|
|
||||||
commands_tx: mpsc::Sender<Command>,
|
|
||||||
}
|
|
||||||
|
|
||||||
enum Command {
|
|
||||||
GetEncoded {
|
|
||||||
chunk: ChunkPosition,
|
|
||||||
reply: oneshot::Sender<Option<Arc<[u8]>>>,
|
|
||||||
},
|
|
||||||
|
|
||||||
Invalidate {
|
|
||||||
chunk: ChunkPosition,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
impl ChunkEncoder {
|
|
||||||
pub fn start(wall: Arc<Wall>) -> Self {
|
|
||||||
let (commands_tx, commands_rx) = mpsc::channel(32);
|
|
||||||
|
|
||||||
tokio::spawn(Self::service(wall, commands_rx));
|
|
||||||
|
|
||||||
Self { commands_tx }
|
|
||||||
}
|
|
||||||
|
|
||||||
pub async fn encoded(&self, chunk: ChunkPosition) -> Option<Arc<[u8]>> {
|
|
||||||
let (tx, rx) = oneshot::channel();
|
|
||||||
self.commands_tx
|
|
||||||
.send(Command::GetEncoded { chunk, reply: tx })
|
|
||||||
.await
|
|
||||||
.ok()?;
|
|
||||||
rx.await.ok().flatten()
|
|
||||||
}
|
|
||||||
|
|
||||||
pub async fn invalidate(&self, chunk: ChunkPosition) {
|
|
||||||
_ = self.commands_tx.send(Command::Invalidate { chunk }).await;
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn invalidate_blocking(&self, chunk: ChunkPosition) {
|
|
||||||
_ = self
|
|
||||||
.commands_tx
|
|
||||||
.blocking_send(Command::Invalidate { chunk });
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn encode(wall: &Wall, chunk: ChunkPosition) -> Option<Arc<[u8]>> {
|
|
||||||
let pixmap = {
|
|
||||||
// Clone out the pixmap to avoid unnecessary chunk mutex contention while the
|
|
||||||
// chunk is being encoded.
|
|
||||||
let chunk_ref = wall.get_chunk(chunk)?;
|
|
||||||
let chunk = chunk_ref.lock().await;
|
|
||||||
chunk.pixmap.clone()
|
|
||||||
};
|
|
||||||
|
|
||||||
let image = tokio::task::spawn_blocking(move || {
|
|
||||||
let webp = webp::Encoder::new(
|
|
||||||
pixmap.data(),
|
|
||||||
webp::PixelLayout::Rgba,
|
|
||||||
pixmap.width(),
|
|
||||||
pixmap.height(),
|
|
||||||
);
|
|
||||||
// NOTE: There's an unnecessary copy here. Wonder if that kills performance much.
|
|
||||||
webp.encode_lossless().to_vec()
|
|
||||||
})
|
|
||||||
.await
|
|
||||||
.ok()?;
|
|
||||||
|
|
||||||
Some(Arc::from(image))
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn service(wall: Arc<Wall>, mut commands_rx: mpsc::Receiver<Command>) {
|
|
||||||
let mut encoded_lru: IndexMap<ChunkPosition, Option<Arc<[u8]>>> = IndexMap::new();
|
|
||||||
|
|
||||||
while let Some(command) = commands_rx.recv().await {
|
|
||||||
match command {
|
|
||||||
Command::GetEncoded { chunk, reply } => {
|
|
||||||
if let Some(encoded) = encoded_lru.get(&chunk) {
|
|
||||||
_ = reply.send(encoded.clone())
|
|
||||||
} else {
|
|
||||||
let encoded = Self::encode(&wall, chunk).await;
|
|
||||||
// TODO: Make this capacity configurable.
|
|
||||||
// 598 is chosen because under the default configuration, it would
|
|
||||||
// correspond to roughly two 3840x2160 displays.
|
|
||||||
if encoded_lru.len() >= 598 {
|
|
||||||
encoded_lru.shift_remove_index(0);
|
|
||||||
}
|
|
||||||
encoded_lru.insert(chunk, encoded.clone());
|
|
||||||
_ = reply.send(encoded);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
Command::Invalidate { chunk } => {
|
|
||||||
encoded_lru.shift_remove(&chunk);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
238
crates/rkgk/src/wall/chunk_images.rs
Normal file
238
crates/rkgk/src/wall/chunk_images.rs
Normal file
|
@ -0,0 +1,238 @@
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
|
use dashmap::DashSet;
|
||||||
|
use eyre::Context;
|
||||||
|
use haku::render::tiny_skia::{IntSize, Pixmap};
|
||||||
|
use rayon::iter::{IntoParallelIterator, IntoParallelRefIterator, ParallelIterator};
|
||||||
|
use tokio::sync::{mpsc, oneshot};
|
||||||
|
use tracing::{error, info, instrument};
|
||||||
|
|
||||||
|
use super::{database::ChunkDataPair, ChunkPosition, Database, Wall};
|
||||||
|
|
||||||
|
/// Chunk image encoding, caching, and storage service.
|
||||||
|
pub struct ChunkImages {
|
||||||
|
wall: Arc<Wall>,
|
||||||
|
async_loop: Arc<ChunkImageLoop>,
|
||||||
|
commands_tx: mpsc::Sender<Command>,
|
||||||
|
}
|
||||||
|
|
||||||
|
enum Command {
|
||||||
|
Encode {
|
||||||
|
chunks: Vec<ChunkPosition>,
|
||||||
|
reply: oneshot::Sender<Vec<ChunkDataPair>>,
|
||||||
|
},
|
||||||
|
|
||||||
|
Load {
|
||||||
|
chunks: Vec<ChunkPosition>,
|
||||||
|
reply: oneshot::Sender<eyre::Result<()>>,
|
||||||
|
},
|
||||||
|
|
||||||
|
MarkModified {
|
||||||
|
chunks: Vec<ChunkPosition>,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
impl ChunkImages {
|
||||||
|
pub fn new(wall: Arc<Wall>, db: Arc<Database>) -> Self {
|
||||||
|
let (commands_tx, commands_rx) = mpsc::channel(32);
|
||||||
|
|
||||||
|
let async_loop = Arc::new(ChunkImageLoop {
|
||||||
|
wall: Arc::clone(&wall),
|
||||||
|
db,
|
||||||
|
chunks_in_db: DashSet::new(),
|
||||||
|
});
|
||||||
|
tokio::spawn(Arc::clone(&async_loop).enter(commands_rx));
|
||||||
|
|
||||||
|
Self {
|
||||||
|
wall,
|
||||||
|
async_loop,
|
||||||
|
commands_tx,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn encoded(&self, chunks: Vec<ChunkPosition>) -> Vec<ChunkDataPair> {
|
||||||
|
let (tx, rx) = oneshot::channel();
|
||||||
|
_ = self
|
||||||
|
.commands_tx
|
||||||
|
.send(Command::Encode { chunks, reply: tx })
|
||||||
|
.await
|
||||||
|
.ok();
|
||||||
|
rx.await.ok().unwrap_or_default()
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn load(&self, chunks: Vec<ChunkPosition>) -> eyre::Result<()> {
|
||||||
|
let (tx, rx) = oneshot::channel();
|
||||||
|
self.commands_tx
|
||||||
|
.send(Command::Load { chunks, reply: tx })
|
||||||
|
.await
|
||||||
|
.context("database is offline")?;
|
||||||
|
rx.await.context("failed to load chunks")?
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn mark_modified_blocking(&self, chunks: Vec<ChunkPosition>) {
|
||||||
|
_ = self
|
||||||
|
.commands_tx
|
||||||
|
.blocking_send(Command::MarkModified { chunks });
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn chunk_exists(&self, position: ChunkPosition) -> bool {
|
||||||
|
self.wall.has_chunk(position) || self.async_loop.chunks_in_db.contains(&position)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
struct ChunkImageLoop {
|
||||||
|
wall: Arc<Wall>,
|
||||||
|
db: Arc<Database>,
|
||||||
|
chunks_in_db: DashSet<ChunkPosition>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl ChunkImageLoop {
|
||||||
|
#[instrument(skip(self, reply))]
|
||||||
|
async fn encode(
|
||||||
|
self: Arc<Self>,
|
||||||
|
mut chunks: Vec<ChunkPosition>,
|
||||||
|
reply: oneshot::Sender<Vec<ChunkDataPair>>,
|
||||||
|
) {
|
||||||
|
// Any chunks that are _not_ loaded, we will send back directly from the database.
|
||||||
|
let mut unloaded = vec![];
|
||||||
|
chunks.retain(|&position| {
|
||||||
|
let is_loaded = self.wall.has_chunk(position);
|
||||||
|
unloaded.push(position);
|
||||||
|
is_loaded
|
||||||
|
});
|
||||||
|
|
||||||
|
// Any chunks that _are_ loaded, we will encode into WebP.
|
||||||
|
let this = Arc::clone(&self);
|
||||||
|
let loaded = tokio::task::spawn_blocking(move || {
|
||||||
|
chunks
|
||||||
|
.into_par_iter()
|
||||||
|
.flat_map(|position| -> Option<_> {
|
||||||
|
let pixmap = {
|
||||||
|
// Clone out the pixmap to avoid unnecessary chunk mutex contention while the
|
||||||
|
// chunk is being encoded.
|
||||||
|
let chunk_ref = this.wall.get_chunk(position)?;
|
||||||
|
let chunk = chunk_ref.blocking_lock();
|
||||||
|
chunk.pixmap.clone()
|
||||||
|
};
|
||||||
|
|
||||||
|
let webp = webp::Encoder::new(
|
||||||
|
pixmap.data(),
|
||||||
|
webp::PixelLayout::Rgba,
|
||||||
|
pixmap.width(),
|
||||||
|
pixmap.height(),
|
||||||
|
);
|
||||||
|
// NOTE: There's an unnecessary copy here. Wonder if that kills performance much.
|
||||||
|
Some(ChunkDataPair {
|
||||||
|
position,
|
||||||
|
data: Arc::from(webp.encode_lossless().to_vec()),
|
||||||
|
})
|
||||||
|
})
|
||||||
|
.collect::<Vec<_>>()
|
||||||
|
})
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
// We'll also write the loaded chunks back to the database while we have the images.
|
||||||
|
// No point wasting the encoding time.
|
||||||
|
if !loaded.is_empty() {
|
||||||
|
info!(num = loaded.len(), "writing loaded chunks to database");
|
||||||
|
}
|
||||||
|
_ = self.db.write_chunks(loaded.clone()).await;
|
||||||
|
for &ChunkDataPair { position, .. } in &loaded {
|
||||||
|
self.chunks_in_db.insert(position);
|
||||||
|
}
|
||||||
|
|
||||||
|
let mut all = loaded;
|
||||||
|
match self.db.read_chunks(unloaded).await {
|
||||||
|
Ok(mut chunks) => all.append(&mut chunks),
|
||||||
|
Err(err) => error!(?err, "read_chunks failed to read unloaded chunks"),
|
||||||
|
}
|
||||||
|
|
||||||
|
_ = reply.send(all);
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn load_inner(self: Arc<Self>, mut chunks: Vec<ChunkPosition>) -> eyre::Result<()> {
|
||||||
|
// Skip already loaded chunks.
|
||||||
|
chunks.retain(|&position| !self.wall.has_chunk(position));
|
||||||
|
if chunks.is_empty() {
|
||||||
|
return Ok(());
|
||||||
|
}
|
||||||
|
|
||||||
|
info!(?chunks, "to load");
|
||||||
|
|
||||||
|
let chunks = self.db.read_chunks(chunks.clone()).await?;
|
||||||
|
|
||||||
|
let chunks2 = chunks.clone();
|
||||||
|
let decoded = tokio::task::spawn_blocking(move || {
|
||||||
|
chunks2
|
||||||
|
.par_iter()
|
||||||
|
.flat_map(|ChunkDataPair { position, data }| {
|
||||||
|
webp::Decoder::new(data)
|
||||||
|
.decode()
|
||||||
|
.and_then(|image| {
|
||||||
|
info!(?position, "decoded");
|
||||||
|
let size = IntSize::from_wh(image.width(), image.height())?;
|
||||||
|
Pixmap::from_vec(image.to_vec(), size)
|
||||||
|
})
|
||||||
|
.map(|pixmap| (*position, pixmap))
|
||||||
|
})
|
||||||
|
.collect::<Vec<(ChunkPosition, Pixmap)>>()
|
||||||
|
})
|
||||||
|
.await
|
||||||
|
.context("failed to decode chunks from the database")?;
|
||||||
|
|
||||||
|
// I don't know yet if locking all the chunks is a good idea at this point.
|
||||||
|
// I can imagine contended chunks having some trouble loading.
|
||||||
|
let chunk_arcs: Vec<_> = chunks
|
||||||
|
.iter()
|
||||||
|
.map(|ChunkDataPair { position, .. }| self.wall.get_or_create_chunk(*position))
|
||||||
|
.collect();
|
||||||
|
let mut chunk_refs = Vec::with_capacity(chunk_arcs.len());
|
||||||
|
for arc in &chunk_arcs {
|
||||||
|
chunk_refs.push(arc.lock().await);
|
||||||
|
}
|
||||||
|
|
||||||
|
for ((_, pixmap), mut chunk) in decoded.into_iter().zip(chunk_refs) {
|
||||||
|
chunk.pixmap = pixmap;
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[instrument(skip(self, reply))]
|
||||||
|
async fn load(
|
||||||
|
self: Arc<Self>,
|
||||||
|
chunks: Vec<ChunkPosition>,
|
||||||
|
reply: oneshot::Sender<eyre::Result<()>>,
|
||||||
|
) {
|
||||||
|
_ = reply.send(self.load_inner(chunks).await);
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn enter(self: Arc<Self>, mut commands_rx: mpsc::Receiver<Command>) {
|
||||||
|
let all_chunks = self
|
||||||
|
.db
|
||||||
|
.get_all_chunks()
|
||||||
|
.await
|
||||||
|
.expect("could not list chunks in the database");
|
||||||
|
for position in all_chunks {
|
||||||
|
self.chunks_in_db.insert(position);
|
||||||
|
}
|
||||||
|
|
||||||
|
while let Some(command) = commands_rx.recv().await {
|
||||||
|
match command {
|
||||||
|
Command::Encode { chunks, reply } => {
|
||||||
|
// TODO: This should have a caching layer.
|
||||||
|
tokio::spawn(Arc::clone(&self).encode(chunks, reply));
|
||||||
|
}
|
||||||
|
|
||||||
|
Command::Load { chunks, reply } => {
|
||||||
|
tokio::spawn(Arc::clone(&self).load(chunks, reply));
|
||||||
|
}
|
||||||
|
|
||||||
|
Command::MarkModified { chunks } => {
|
||||||
|
// TODO: This should invalidate data from the caching layer.
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -1,3 +1,5 @@
|
||||||
|
use std::iter::Take;
|
||||||
|
|
||||||
use super::ChunkPosition;
|
use super::ChunkPosition;
|
||||||
|
|
||||||
#[derive(Debug, Clone, Copy, PartialEq, Eq)]
|
#[derive(Debug, Clone, Copy, PartialEq, Eq)]
|
||||||
|
@ -8,13 +10,35 @@ pub struct ChunkIterator {
|
||||||
}
|
}
|
||||||
|
|
||||||
impl ChunkIterator {
|
impl ChunkIterator {
|
||||||
pub fn new(top_left: ChunkPosition, bottom_right: ChunkPosition) -> Self {
|
pub fn new(start: ChunkPosition, end: ChunkPosition) -> Self {
|
||||||
|
let top_left = ChunkPosition::new(start.x.min(end.x), start.y.min(end.y));
|
||||||
|
let bottom_right = ChunkPosition::new(start.x.max(end.x), start.y.max(end.y));
|
||||||
Self {
|
Self {
|
||||||
cursor: top_left,
|
cursor: top_left,
|
||||||
left: top_left.x,
|
left: top_left.x,
|
||||||
bottom_right,
|
bottom_right,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn take_next(&mut self, n: i32) -> Take<Self> {
|
||||||
|
assert!(n > 0);
|
||||||
|
|
||||||
|
let take = (*self).take(n as usize);
|
||||||
|
|
||||||
|
let x = self.cursor.x - self.left;
|
||||||
|
let width = self.bottom_right.x - self.left;
|
||||||
|
if width != 0 {
|
||||||
|
self.cursor.x = self.left + (x + n) % width;
|
||||||
|
self.cursor.y += n / width;
|
||||||
|
} else {
|
||||||
|
// In a width = 0 configuration, we iterate vertically.
|
||||||
|
// This is probably not the right thing to do, but we're just doing this to guard
|
||||||
|
// against malicious clients.
|
||||||
|
self.cursor.y += n;
|
||||||
|
}
|
||||||
|
|
||||||
|
take
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Iterator for ChunkIterator {
|
impl Iterator for ChunkIterator {
|
||||||
|
|
324
crates/rkgk/src/wall/database.rs
Normal file
324
crates/rkgk/src/wall/database.rs
Normal file
|
@ -0,0 +1,324 @@
|
||||||
|
use std::{
|
||||||
|
convert::identity,
|
||||||
|
path::{Path, PathBuf},
|
||||||
|
sync::Arc,
|
||||||
|
};
|
||||||
|
|
||||||
|
use chrono::Utc;
|
||||||
|
use eyre::Context;
|
||||||
|
use rusqlite::Connection;
|
||||||
|
use tokio::sync::{mpsc, oneshot};
|
||||||
|
use tracing::{error, info, instrument};
|
||||||
|
|
||||||
|
use crate::login::UserId;
|
||||||
|
|
||||||
|
use super::{ChunkPosition, WallId};
|
||||||
|
|
||||||
|
pub struct Settings {
|
||||||
|
pub path: PathBuf,
|
||||||
|
pub wall_id: WallId,
|
||||||
|
pub default_wall_settings: super::Settings,
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct Database {
|
||||||
|
wall_settings: super::Settings,
|
||||||
|
command_tx: mpsc::Sender<Command>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone)]
|
||||||
|
pub struct ChunkDataPair {
|
||||||
|
pub position: ChunkPosition,
|
||||||
|
pub data: Arc<[u8]>,
|
||||||
|
}
|
||||||
|
|
||||||
|
enum Command {
|
||||||
|
SetWallInfo {
|
||||||
|
created_by: UserId,
|
||||||
|
title: String,
|
||||||
|
reply: oneshot::Sender<eyre::Result<()>>,
|
||||||
|
},
|
||||||
|
|
||||||
|
WriteChunks {
|
||||||
|
chunks: Vec<ChunkDataPair>,
|
||||||
|
reply: oneshot::Sender<eyre::Result<()>>,
|
||||||
|
},
|
||||||
|
|
||||||
|
ReadChunks {
|
||||||
|
chunks: Vec<ChunkPosition>,
|
||||||
|
reply: oneshot::Sender<Vec<ChunkDataPair>>,
|
||||||
|
},
|
||||||
|
|
||||||
|
GetAllChunks {
|
||||||
|
reply: oneshot::Sender<eyre::Result<Vec<ChunkPosition>>>,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Database {
|
||||||
|
pub fn wall_settings(&self) -> &super::Settings {
|
||||||
|
&self.wall_settings
|
||||||
|
}
|
||||||
|
|
||||||
|
#[instrument(skip(self), err(Debug))]
|
||||||
|
pub async fn write_chunks(&self, chunks: Vec<ChunkDataPair>) -> eyre::Result<()> {
|
||||||
|
let (tx, rx) = oneshot::channel();
|
||||||
|
self.command_tx
|
||||||
|
.send(Command::WriteChunks { chunks, reply: tx })
|
||||||
|
.await
|
||||||
|
.context("database is offline")?;
|
||||||
|
rx.await.context("database returned an error")?
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn read_chunks(
|
||||||
|
&self,
|
||||||
|
chunks: Vec<ChunkPosition>,
|
||||||
|
) -> eyre::Result<Vec<ChunkDataPair>> {
|
||||||
|
let (tx, rx) = oneshot::channel();
|
||||||
|
self.command_tx
|
||||||
|
.send(Command::ReadChunks { chunks, reply: tx })
|
||||||
|
.await
|
||||||
|
.context("database is offline")?;
|
||||||
|
rx.await.context("database did not return anything")
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn get_all_chunks(&self) -> eyre::Result<Vec<ChunkPosition>> {
|
||||||
|
let (tx, rx) = oneshot::channel();
|
||||||
|
self.command_tx
|
||||||
|
.send(Command::GetAllChunks { reply: tx })
|
||||||
|
.await
|
||||||
|
.context("database is offline")?;
|
||||||
|
rx.await.context("database did not return anything")?
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[instrument(name = "wall::database::start", skip(settings), fields(wall_id = %settings.wall_id))]
|
||||||
|
pub fn start(settings: Settings) -> eyre::Result<Database> {
|
||||||
|
let db = Connection::open(settings.path).context("cannot open wall database")?;
|
||||||
|
|
||||||
|
let major: u32 = env!("CARGO_PKG_VERSION_MAJOR").parse().unwrap();
|
||||||
|
let minor: u32 = env!("CARGO_PKG_VERSION_MINOR").parse().unwrap();
|
||||||
|
let patch: u32 = env!("CARGO_PKG_VERSION_PATCH").parse().unwrap();
|
||||||
|
let version = major * 1_000_000 + minor * 1_000 + patch;
|
||||||
|
|
||||||
|
info!("initial setup");
|
||||||
|
|
||||||
|
db.execute_batch(
|
||||||
|
r#"
|
||||||
|
PRAGMA application_id = 0x726B6757; -- rkgW
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS
|
||||||
|
t_file_info (
|
||||||
|
id INTEGER PRIMARY KEY CHECK (id = 1),
|
||||||
|
version INTEGER NOT NULL,
|
||||||
|
wall_id BLOB NOT NULL,
|
||||||
|
mtime INTEGER NOT NULL DEFAULT (unixepoch())
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS
|
||||||
|
t_wall_settings (
|
||||||
|
id INTEGER PRIMARY KEY CHECK (id = 1),
|
||||||
|
max_chunks INTEGER NOT NULL,
|
||||||
|
max_sessions INTEGER NOT NULL,
|
||||||
|
paint_area INTEGER NOT NULL,
|
||||||
|
chunk_size INTEGER NOT NULL
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS
|
||||||
|
t_wall_info (
|
||||||
|
id INTEGER PRIMARY KEY CHECK (id = 1),
|
||||||
|
created_by BLOB NOT NULL,
|
||||||
|
title TEXT NOT NULL
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS
|
||||||
|
t_chunks_v0 (
|
||||||
|
chunk_index INTEGER PRIMARY KEY,
|
||||||
|
chunk_x INTEGER NOT NULL,
|
||||||
|
chunk_y INTEGER NOT NULL,
|
||||||
|
mtime INTEGER NOT NULL DEFAULT (unixepoch()),
|
||||||
|
webp_data BLOB NOT NULL,
|
||||||
|
|
||||||
|
UNIQUE(chunk_x, chunk_y) ON CONFLICT REPLACE
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE INDEX IF NOT EXISTS
|
||||||
|
t_chunks_v0_index_xy ON t_chunks_v0
|
||||||
|
(chunk_x, chunk_y);
|
||||||
|
"#,
|
||||||
|
)?;
|
||||||
|
|
||||||
|
info!("set file version and wall ID");
|
||||||
|
|
||||||
|
db.execute(
|
||||||
|
r#"
|
||||||
|
INSERT OR IGNORE
|
||||||
|
INTO t_file_info
|
||||||
|
(version, wall_id, mtime)
|
||||||
|
VALUES (?, ?, unixepoch());
|
||||||
|
"#,
|
||||||
|
(version, settings.wall_id.0),
|
||||||
|
)?;
|
||||||
|
|
||||||
|
info!("set wall mtime");
|
||||||
|
|
||||||
|
db.execute(
|
||||||
|
r#"
|
||||||
|
UPDATE t_file_info
|
||||||
|
SET mtime = unixepoch();
|
||||||
|
"#,
|
||||||
|
(),
|
||||||
|
)?;
|
||||||
|
|
||||||
|
info!("initialize/get wall settings");
|
||||||
|
|
||||||
|
db.execute(
|
||||||
|
r#"
|
||||||
|
INSERT OR IGNORE
|
||||||
|
INTO t_wall_settings
|
||||||
|
(max_chunks, max_sessions, paint_area, chunk_size)
|
||||||
|
VALUES (?, ?, ?, ?);
|
||||||
|
"#,
|
||||||
|
(
|
||||||
|
settings.default_wall_settings.max_chunks,
|
||||||
|
settings.default_wall_settings.max_sessions,
|
||||||
|
settings.default_wall_settings.paint_area,
|
||||||
|
settings.default_wall_settings.chunk_size,
|
||||||
|
),
|
||||||
|
)?;
|
||||||
|
|
||||||
|
let wall_settings = db.query_row(
|
||||||
|
r#"
|
||||||
|
SELECT
|
||||||
|
max_chunks, max_sessions, paint_area, chunk_size
|
||||||
|
FROM t_wall_settings;
|
||||||
|
"#,
|
||||||
|
(),
|
||||||
|
|row| {
|
||||||
|
Ok(super::Settings {
|
||||||
|
max_chunks: row.get(0)?,
|
||||||
|
max_sessions: row.get(1)?,
|
||||||
|
paint_area: row.get(2)?,
|
||||||
|
chunk_size: row.get(3)?,
|
||||||
|
})
|
||||||
|
},
|
||||||
|
)?;
|
||||||
|
|
||||||
|
let (command_tx, mut command_rx) = mpsc::channel(8);
|
||||||
|
|
||||||
|
std::thread::Builder::new()
|
||||||
|
.name(format!("database thread {}", settings.wall_id))
|
||||||
|
.spawn(move || {
|
||||||
|
let mut s_set_wall_info = db
|
||||||
|
.prepare(
|
||||||
|
r#"
|
||||||
|
INSERT OR REPLACE
|
||||||
|
INTO t_wall_info
|
||||||
|
(created_by, title)
|
||||||
|
VALUES (?, ?);
|
||||||
|
"#,
|
||||||
|
)
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
let mut s_write_chunk = db
|
||||||
|
.prepare(
|
||||||
|
r#"
|
||||||
|
INSERT
|
||||||
|
INTO t_chunks_v0
|
||||||
|
(chunk_x, chunk_y, webp_data, mtime)
|
||||||
|
VALUES (?, ?, ?, unixepoch());
|
||||||
|
"#,
|
||||||
|
)
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
let mut s_read_chunk = db
|
||||||
|
.prepare(
|
||||||
|
r#"
|
||||||
|
SELECT webp_data
|
||||||
|
FROM t_chunks_v0
|
||||||
|
WHERE chunk_x = ? AND chunk_y = ?;
|
||||||
|
"#,
|
||||||
|
)
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
let mut s_get_all_chunks = db
|
||||||
|
.prepare(
|
||||||
|
r#"
|
||||||
|
SELECT chunk_x, chunk_y
|
||||||
|
FROM t_chunks_v0;
|
||||||
|
"#,
|
||||||
|
)
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
while let Some(command) = command_rx.blocking_recv() {
|
||||||
|
match command {
|
||||||
|
Command::SetWallInfo {
|
||||||
|
created_by,
|
||||||
|
title,
|
||||||
|
reply,
|
||||||
|
} => {
|
||||||
|
_ = reply.send(
|
||||||
|
s_set_wall_info
|
||||||
|
.execute((created_by.0, title))
|
||||||
|
.map(|_| ())
|
||||||
|
.context("failed to set wall info"),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
Command::WriteChunks { chunks, reply } => {
|
||||||
|
let mut result = Ok(());
|
||||||
|
for ChunkDataPair { position, data } in chunks {
|
||||||
|
if let Err(error) =
|
||||||
|
s_write_chunk.execute((position.x, position.y, &data[..]))
|
||||||
|
{
|
||||||
|
result = Err(error).with_context(|| {
|
||||||
|
format!("failed to update chunk at {position:?}")
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
_ = reply.send(result.context(
|
||||||
|
"failed to update one or more chunks; see context for last error",
|
||||||
|
));
|
||||||
|
}
|
||||||
|
|
||||||
|
Command::ReadChunks { chunks, reply } => {
|
||||||
|
let result = chunks
|
||||||
|
.into_iter()
|
||||||
|
.flat_map(|position| {
|
||||||
|
s_read_chunk
|
||||||
|
.query_row((position.x, position.y), |row| {
|
||||||
|
Ok(ChunkDataPair {
|
||||||
|
position,
|
||||||
|
data: Arc::from(row.get::<_, Vec<u8>>(0)?),
|
||||||
|
})
|
||||||
|
})
|
||||||
|
.inspect_err(|err| {
|
||||||
|
if err != &rusqlite::Error::QueryReturnedNoRows {
|
||||||
|
error!(?err, ?position, "while reading chunk");
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.ok()
|
||||||
|
})
|
||||||
|
.collect();
|
||||||
|
_ = reply.send(result);
|
||||||
|
}
|
||||||
|
|
||||||
|
Command::GetAllChunks { reply } => {
|
||||||
|
_ = reply.send(
|
||||||
|
s_get_all_chunks
|
||||||
|
.query_map((), |row| {
|
||||||
|
Ok(ChunkPosition::new(row.get(0)?, row.get(1)?))
|
||||||
|
})
|
||||||
|
.map(|chunks| chunks.collect::<Result<_, _>>())
|
||||||
|
.and_then(identity)
|
||||||
|
.context("failed to query all chunks"),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.context("cannot spawn thread")?;
|
||||||
|
|
||||||
|
Ok(Database {
|
||||||
|
command_tx,
|
||||||
|
wall_settings,
|
||||||
|
})
|
||||||
|
}
|
10
rkgk.toml
10
rkgk.toml
|
@ -1,4 +1,4 @@
|
||||||
[wall]
|
[wall_broker.default_wall_settings]
|
||||||
|
|
||||||
# The settings below control the creation of new walls.
|
# The settings below control the creation of new walls.
|
||||||
|
|
||||||
|
@ -13,7 +13,8 @@ max_sessions = 128
|
||||||
|
|
||||||
# The size of chunks.
|
# The size of chunks.
|
||||||
# Choosing an appropriate size for chunks is a tradeoff between performance and disk space - 168 is
|
# Choosing an appropriate size for chunks is a tradeoff between performance and disk space - 168 is
|
||||||
# chosen as a reasonable default
|
# chosen as a reasonable default which is just small enough to perform operations on fairly quickly
|
||||||
|
# and responsively.
|
||||||
chunk_size = 168
|
chunk_size = 168
|
||||||
|
|
||||||
# The size of the area that can be drawn over by a brush, in pixels.
|
# The size of the area that can be drawn over by a brush, in pixels.
|
||||||
|
@ -21,6 +22,11 @@ chunk_size = 168
|
||||||
# can produce.
|
# can produce.
|
||||||
paint_area = 504
|
paint_area = 504
|
||||||
|
|
||||||
|
[wall_broker.auto_save]
|
||||||
|
|
||||||
|
# How often should modified chunks be saved to the database.
|
||||||
|
interval_seconds = 10
|
||||||
|
|
||||||
[haku]
|
[haku]
|
||||||
|
|
||||||
# The settings below control the Haku runtime on the server side.
|
# The settings below control the Haku runtime on the server side.
|
||||||
|
|
|
@ -90,10 +90,6 @@ class CanvasRenderer extends HTMLElement {
|
||||||
this.ctx.globalCompositeOperation = "source-over";
|
this.ctx.globalCompositeOperation = "source-over";
|
||||||
this.ctx.drawImage(chunk.canvas, x, y);
|
this.ctx.drawImage(chunk.canvas, x, y);
|
||||||
}
|
}
|
||||||
|
|
||||||
this.ctx.globalCompositeOperation = "difference";
|
|
||||||
this.ctx.fillStyle = "white";
|
|
||||||
this.ctx.fillText(`${chunkX}, ${chunkY}`, x, y + 12);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -86,7 +86,6 @@ reticleRenderer.connectViewport(canvasRenderer.viewport);
|
||||||
let sendViewportUpdate = debounce(updateInterval, () => {
|
let sendViewportUpdate = debounce(updateInterval, () => {
|
||||||
let visibleRect = canvasRenderer.getVisibleChunkRect();
|
let visibleRect = canvasRenderer.getVisibleChunkRect();
|
||||||
session.sendViewport(visibleRect);
|
session.sendViewport(visibleRect);
|
||||||
console.log("visibleRect", visibleRect);
|
|
||||||
});
|
});
|
||||||
canvasRenderer.addEventListener(".viewportUpdate", sendViewportUpdate);
|
canvasRenderer.addEventListener(".viewportUpdate", sendViewportUpdate);
|
||||||
sendViewportUpdate();
|
sendViewportUpdate();
|
||||||
|
|
Loading…
Reference in a new issue