Compare commits
2 commits
c8b8b56456
...
dc0fbe6242
Author | SHA1 | Date | |
---|---|---|---|
dc0fbe6242 | |||
5f598ece03 |
5 changed files with 309 additions and 88 deletions
44
default.nix
44
default.nix
|
@ -88,7 +88,7 @@ lib.makeScope pkgs.newScope (
|
||||||
datasette-settings = self.callPackage (
|
datasette-settings = self.callPackage (
|
||||||
{ formats }:
|
{ formats }:
|
||||||
(formats.json { }).generate "datasette-settings.json" {
|
(formats.json { }).generate "datasette-settings.json" {
|
||||||
sql_time_limit_ms = 8000;
|
sql_time_limit_ms = 16000;
|
||||||
}
|
}
|
||||||
) { };
|
) { };
|
||||||
datasette-metadata = self.callPackage (
|
datasette-metadata = self.callPackage (
|
||||||
|
@ -122,38 +122,18 @@ lib.makeScope pkgs.newScope (
|
||||||
<code>CudaArtifact</code>s (identified by <code>sha256</code>)
|
<code>CudaArtifact</code>s (identified by <code>sha256</code>)
|
||||||
claiming the same <code>(pname, version, platform)</code> triple
|
claiming the same <code>(pname, version, platform)</code> triple
|
||||||
'';
|
'';
|
||||||
sql = ''
|
sql = builtins.readFile ./src/q/summary-cuda-conflicts.sql;
|
||||||
SELECT
|
};
|
||||||
COUNT(DISTINCT sha256) AS conflicts,
|
queries.cuda_pnames = {
|
||||||
pname.str AS pname,
|
title = "Known CUDA Artifacts";
|
||||||
ver.str AS ver,
|
description_html = ''
|
||||||
plat.str AS plat,
|
Overview of known CUDA artifacts sorted by <code>pname</code>
|
||||||
GROUP_CONCAT(name.str, char(10)) AS name,
|
|
||||||
GROUP_CONCAT(tag.str, char(10)) AS tag,
|
|
||||||
GROUP_CONCAT(h.hash, char(10)) AS sha256
|
|
||||||
FROM
|
|
||||||
(
|
|
||||||
CudaArtifact AS cc,
|
|
||||||
Str AS name,
|
|
||||||
Str AS pname,
|
|
||||||
Str as ver,
|
|
||||||
Str as plat,
|
|
||||||
Hash as h
|
|
||||||
ON cc.name=name.id
|
|
||||||
AND cc.pname=pname.id
|
|
||||||
AND cc.version = ver.id
|
|
||||||
AND cc.platform = plat.id
|
|
||||||
AND cc.sha256 = h.id
|
|
||||||
)
|
|
||||||
LEFT JOIN Str AS tag
|
|
||||||
ON
|
|
||||||
cc.compat_tag=tag.id
|
|
||||||
GROUP BY
|
|
||||||
cc.pname, cc.version, cc.platform
|
|
||||||
HAVING
|
|
||||||
conflicts >= CAST(:min_conflicts AS INTEGER)
|
|
||||||
ORDER BY conflicts DESC
|
|
||||||
'';
|
'';
|
||||||
|
sql = builtins.readFile ./src/q/summary-cuda-pnames.sql;
|
||||||
|
};
|
||||||
|
queries.cuda_platforms = {
|
||||||
|
title = "CUDA: Supported Platforms ";
|
||||||
|
sql = builtins.readFile ./src/q/summary-cuda-platforms.sql;
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
159
src/main.rs
159
src/main.rs
|
@ -186,7 +186,7 @@ async fn open_context(
|
||||||
}
|
}
|
||||||
|
|
||||||
let con = rusqlite::Connection::open(&db_path).expect("Failed to construct Database object");
|
let con = rusqlite::Connection::open(&db_path).expect("Failed to construct Database object");
|
||||||
con.pragma_update(None, "jorunal_mode", "wal").unwrap();
|
con.pragma_update(None, "journal_mode", "wal").unwrap();
|
||||||
con.pragma_update(None, "synchronous", "normal").unwrap();
|
con.pragma_update(None, "synchronous", "normal").unwrap();
|
||||||
con.pragma_update(None, "temp_store", "memory").unwrap();
|
con.pragma_update(None, "temp_store", "memory").unwrap();
|
||||||
con.pragma_update(None, "foreign_keys", "on").unwrap();
|
con.pragma_update(None, "foreign_keys", "on").unwrap();
|
||||||
|
@ -238,6 +238,86 @@ where
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
trait ConnectionLike {
|
||||||
|
#[allow(dead_code)]
|
||||||
|
fn execute<P: rusqlite::Params>(&self, sql: &str, params: P) -> rusqlite::Result<usize>;
|
||||||
|
fn prepare_cached(&self, sql: &str) -> rusqlite::Result<rusqlite::CachedStatement<'_>>;
|
||||||
|
}
|
||||||
|
|
||||||
|
impl ConnectionLike for rusqlite::Connection {
|
||||||
|
fn execute<P: rusqlite::Params>(&self, sql: &str, params: P) -> rusqlite::Result<usize> {
|
||||||
|
<rusqlite::Connection>::execute(self, sql, params)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn prepare_cached(&self, sql: &str) -> rusqlite::Result<rusqlite::CachedStatement<'_>> {
|
||||||
|
<rusqlite::Connection>::prepare_cached(self, sql)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a> ConnectionLike for rusqlite::Transaction<'a> {
|
||||||
|
fn execute<P: rusqlite::Params>(&self, sql: &str, params: P) -> rusqlite::Result<usize> {
|
||||||
|
<rusqlite::Connection>::execute(self, sql, params)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn prepare_cached(&self, sql: &str) -> rusqlite::Result<rusqlite::CachedStatement<'_>> {
|
||||||
|
<rusqlite::Connection>::prepare_cached(self, sql)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a> ConnectionLike for rusqlite::Savepoint<'a> {
|
||||||
|
fn execute<P: rusqlite::Params>(&self, sql: &str, params: P) -> rusqlite::Result<usize> {
|
||||||
|
<rusqlite::Connection>::execute(self, sql, params)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn prepare_cached(&self, sql: &str) -> rusqlite::Result<rusqlite::CachedStatement<'_>> {
|
||||||
|
<rusqlite::Connection>::prepare_cached(self, sql)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn db_add_sample(
|
||||||
|
tx: &mut rusqlite::Transaction<'_>,
|
||||||
|
uri: &str,
|
||||||
|
hash: &Option<String>,
|
||||||
|
http_code: &Option<u16>,
|
||||||
|
content_type: &Option<String>,
|
||||||
|
) -> Result<(u32, u32), Error> {
|
||||||
|
let mut sp = tx.savepoint().context("db_add_sample")?;
|
||||||
|
sp.set_drop_behavior(rusqlite::DropBehavior::Commit);
|
||||||
|
{
|
||||||
|
if let Some(h) = hash {
|
||||||
|
db_add_blob(&mut sp, &h.clone(), None)?;
|
||||||
|
};
|
||||||
|
let mut add_sample = sp
|
||||||
|
.prepare_cached(include_str!("q/add-sample.sql"))
|
||||||
|
.context("Failed to prepare add-sample.sql")?;
|
||||||
|
return Ok(add_sample.query_row(
|
||||||
|
named_params! {
|
||||||
|
":uri": uri,
|
||||||
|
":hash": hash,
|
||||||
|
":http_code": http_code,
|
||||||
|
":content_type": content_type
|
||||||
|
},
|
||||||
|
|row| <(u32, u32)>::try_from(row),
|
||||||
|
)?);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
fn db_add_blob<Con: ConnectionLike>(
|
||||||
|
con: &mut Con,
|
||||||
|
hash: &str,
|
||||||
|
n_bytes: Option<u64>,
|
||||||
|
) -> Result<usize, Error> {
|
||||||
|
let mut add_blob = con
|
||||||
|
.prepare_cached(include_str!("q/upsert-blob.sql"))
|
||||||
|
.context("Failed to prepare upsert-blob.sql")?;
|
||||||
|
Ok(add_blob.execute(params![hash, n_bytes,])?)
|
||||||
|
}
|
||||||
|
fn db_add_uri<Con: ConnectionLike>(con: &mut Con, uri: &str) -> Result<usize, Error> {
|
||||||
|
let mut add_uri = con
|
||||||
|
.prepare_cached(include_str!("q/upsert-uri.sql"))
|
||||||
|
.context("Failed to prepare upsert-uri.sql")?;
|
||||||
|
Ok(add_uri.execute(params![uri])?)
|
||||||
|
}
|
||||||
|
|
||||||
impl<BS: BlobService + Clone, DS: DirectoryService + Clone> SidxContext<BS, DS> {
|
impl<BS: BlobService + Clone, DS: DirectoryService + Clone> SidxContext<BS, DS> {
|
||||||
async fn latest_sample(&self, uri: &str) -> Result<Option<Sampled>, Error> {
|
async fn latest_sample(&self, uri: &str) -> Result<Option<Sampled>, Error> {
|
||||||
let lock = self.con.lock().unwrap();
|
let lock = self.con.lock().unwrap();
|
||||||
|
@ -265,42 +345,6 @@ impl<BS: BlobService + Clone, DS: DirectoryService + Clone> SidxContext<BS, DS>
|
||||||
None => Ok(None),
|
None => Ok(None),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
async fn db_add_sample(
|
|
||||||
&self,
|
|
||||||
uri: &str,
|
|
||||||
hash: &Option<String>,
|
|
||||||
http_code: &Option<u16>,
|
|
||||||
content_type: &Option<String>,
|
|
||||||
) -> Result<(u32, u32), Error> {
|
|
||||||
let lock = self.con.lock().expect("Locking mutex for db_add_sample");
|
|
||||||
let mut add_sample = lock
|
|
||||||
.prepare_cached(include_str!("q/add-sample.sql"))
|
|
||||||
.context("Failed to prepare add-sample.sql")?;
|
|
||||||
Ok(add_sample.query_row(
|
|
||||||
named_params! {
|
|
||||||
":uri": uri,
|
|
||||||
":hash": hash,
|
|
||||||
":http_code": http_code,
|
|
||||||
":content_type": content_type
|
|
||||||
},
|
|
||||||
|row| <(u32, u32)>::try_from(row),
|
|
||||||
)?)
|
|
||||||
}
|
|
||||||
async fn db_add_blob(&self, hash: &str, n_bytes: u64) -> Result<usize, Error> {
|
|
||||||
let lock = self.con.lock().expect("db_add_blob: couldn't lock mutex?");
|
|
||||||
let mut add_blob = lock
|
|
||||||
.prepare_cached(include_str!("q/upsert-blob.sql"))
|
|
||||||
.context("Failed to prepare upsert-blob.sql")?;
|
|
||||||
Ok(add_blob.execute(params![hash, n_bytes,])?)
|
|
||||||
}
|
|
||||||
async fn db_add_uri(&self, uri: &str) -> Result<usize, Error> {
|
|
||||||
let lock = self.con.lock().unwrap();
|
|
||||||
let mut add_uri = lock
|
|
||||||
.prepare_cached(include_str!("q/upsert-uri.sql"))
|
|
||||||
.context("Failed to prepare upsert-uri.sql")?;
|
|
||||||
|
|
||||||
Ok(add_uri.execute(params![uri])?)
|
|
||||||
}
|
|
||||||
async fn record_ingested_node(
|
async fn record_ingested_node(
|
||||||
&self,
|
&self,
|
||||||
uri: &str,
|
uri: &str,
|
||||||
|
@ -308,25 +352,28 @@ impl<BS: BlobService + Clone, DS: DirectoryService + Clone> SidxContext<BS, DS>
|
||||||
http_code: Option<u16>,
|
http_code: Option<u16>,
|
||||||
content_type: Option<String>,
|
content_type: Option<String>,
|
||||||
) -> Result<Sampled, Error> {
|
) -> Result<Sampled, Error> {
|
||||||
let digest64 = if let Some(SizedBlob { hash, n_bytes }) = blob {
|
let mut lock = self.con.lock().unwrap();
|
||||||
let digest64 = format!("{}", hash);
|
let mut tx = lock.transaction()?;
|
||||||
self.db_add_blob(&digest64, n_bytes.clone()).await?;
|
{
|
||||||
Some(digest64)
|
let digest64 = if let Some(SizedBlob { hash, n_bytes }) = blob {
|
||||||
} else {
|
let digest64 = format!("{}", hash);
|
||||||
None
|
db_add_blob(&mut tx, &digest64, Some(n_bytes.clone()))?;
|
||||||
};
|
Some(digest64)
|
||||||
self.db_add_uri(&uri).await?;
|
} else {
|
||||||
let (sample_id, epoch) = self
|
None
|
||||||
.db_add_sample(&uri, &digest64, &http_code, &content_type)
|
};
|
||||||
.await?;
|
db_add_uri(&mut tx, &uri)?;
|
||||||
Ok(Sampled {
|
let (sample_id, epoch) =
|
||||||
sample_id,
|
db_add_sample(&mut tx, &uri, &digest64, &http_code, &content_type)?;
|
||||||
uri: uri.to_string(),
|
Ok(Sampled {
|
||||||
blob: blob.clone(),
|
sample_id,
|
||||||
http_status: http_code,
|
uri: uri.to_string(),
|
||||||
epoch,
|
blob: blob.clone(),
|
||||||
when: SampledWhen::Now,
|
http_status: http_code,
|
||||||
})
|
epoch,
|
||||||
|
when: SampledWhen::Now,
|
||||||
|
})
|
||||||
|
}
|
||||||
}
|
}
|
||||||
async fn download(&self, uri: &Url) -> Result<Sampled, Error> {
|
async fn download(&self, uri: &Url) -> Result<Sampled, Error> {
|
||||||
let _permit = self.http_semaphore.acquire().await.unwrap();
|
let _permit = self.http_semaphore.acquire().await.unwrap();
|
||||||
|
|
34
src/q/summary-cuda-conflicts.sql
Normal file
34
src/q/summary-cuda-conflicts.sql
Normal file
|
@ -0,0 +1,34 @@
|
||||||
|
SELECT
|
||||||
|
COUNT(DISTINCT sha256) AS conflicts,
|
||||||
|
pname.str AS pname,
|
||||||
|
ver.str AS ver,
|
||||||
|
plat.str AS plat,
|
||||||
|
GROUP_CONCAT(name.str, char(10)) AS name,
|
||||||
|
GROUP_CONCAT(IFNULL(tag.str, char(0x274C)), char(10)) AS tag,
|
||||||
|
GROUP_CONCAT(h.hash, char(10)) AS sha256
|
||||||
|
FROM
|
||||||
|
(
|
||||||
|
CudaArtifact AS cc,
|
||||||
|
Str AS name,
|
||||||
|
Str AS pname,
|
||||||
|
Str as ver,
|
||||||
|
Str as plat,
|
||||||
|
Hash as h ON cc.name = name.id
|
||||||
|
AND cc.pname = pname.id
|
||||||
|
AND cc.version = ver.id
|
||||||
|
AND cc.platform = plat.id
|
||||||
|
AND cc.sha256 = h.id
|
||||||
|
)
|
||||||
|
LEFT JOIN Str AS tag ON cc.compat_tag = tag.id
|
||||||
|
GROUP BY
|
||||||
|
cc.pname,
|
||||||
|
cc.version,
|
||||||
|
cc.platform
|
||||||
|
HAVING
|
||||||
|
conflicts >= CAST(:min_conflicts AS INTEGER)
|
||||||
|
ORDER BY
|
||||||
|
conflicts DESC,
|
||||||
|
cc.pname,
|
||||||
|
cc.version,
|
||||||
|
cc.platform,
|
||||||
|
tag.str
|
73
src/q/summary-cuda-platforms.sql
Normal file
73
src/q/summary-cuda-platforms.sql
Normal file
|
@ -0,0 +1,73 @@
|
||||||
|
WITH PerQuadruple AS (
|
||||||
|
SELECT
|
||||||
|
pname.str AS pname,
|
||||||
|
IFNULL(ver.str, char(0x274C)) AS version,
|
||||||
|
plat.str AS platform,
|
||||||
|
IFNULL(ct.str, char(0x274C)) AS tag,
|
||||||
|
COUNT(DISTINCT sha256.hash) AS uq_sha256s
|
||||||
|
FROM
|
||||||
|
(
|
||||||
|
CudaArtifact AS ca,
|
||||||
|
Str AS pname,
|
||||||
|
Str AS plat,
|
||||||
|
Hash AS sha256 ON ca.pname = pname.id
|
||||||
|
AND ca.platform = plat.id
|
||||||
|
AND sha256.id = ca.sha256
|
||||||
|
)
|
||||||
|
LEFT JOIN Str AS ver ON ver.id = ca.version
|
||||||
|
LEFT JOIN Str AS ct ON ca.compat_tag = ct.id
|
||||||
|
GROUP BY
|
||||||
|
pname.id,
|
||||||
|
version,
|
||||||
|
platform,
|
||||||
|
tag
|
||||||
|
ORDER BY
|
||||||
|
uq_sha256s DESC,
|
||||||
|
pname.str,
|
||||||
|
version,
|
||||||
|
plat.str,
|
||||||
|
tag
|
||||||
|
),
|
||||||
|
ByPname AS (
|
||||||
|
SELECT
|
||||||
|
pname,
|
||||||
|
REPLACE(GROUP_CONCAT(DISTINCT pq.version), ',', char(10)) AS versions,
|
||||||
|
REPLACE(
|
||||||
|
GROUP_CONCAT(DISTINCT pq.platform),
|
||||||
|
',',
|
||||||
|
char(10)
|
||||||
|
) AS platforms,
|
||||||
|
REPLACE(GROUP_CONCAT(DISTINCT pq.tag), ',', char(10)) AS tags,
|
||||||
|
SUM(pq.uq_sha256s) AS uq_sha256s
|
||||||
|
FROM
|
||||||
|
PerQuadruple AS pq
|
||||||
|
GROUP BY
|
||||||
|
pq.pname
|
||||||
|
),
|
||||||
|
VersionGroups AS (
|
||||||
|
SELECT
|
||||||
|
REPLACE(GROUP_CONCAT(DISTINCT pname), ',', char(10)) AS pnames,
|
||||||
|
versions
|
||||||
|
FROM
|
||||||
|
ByPname AS bp
|
||||||
|
GROUP BY
|
||||||
|
versions
|
||||||
|
ORDER BY
|
||||||
|
LENGTH(pnames) DESC
|
||||||
|
),
|
||||||
|
PlatformGroups AS (
|
||||||
|
SELECT
|
||||||
|
REPLACE(GROUP_CONCAT(DISTINCT pname), ',', char(10)) AS pnames,
|
||||||
|
platforms
|
||||||
|
FROM
|
||||||
|
ByPname AS bp
|
||||||
|
GROUP BY
|
||||||
|
platforms
|
||||||
|
ORDER BY
|
||||||
|
LENGTH(platforms) DESC,
|
||||||
|
LENGTH(pnames)
|
||||||
|
)
|
||||||
|
SELECT
|
||||||
|
*
|
||||||
|
FROM
|
||||||
|
PlatformGroups
|
87
src/q/summary-cuda-pnames.sql
Normal file
87
src/q/summary-cuda-pnames.sql
Normal file
|
@ -0,0 +1,87 @@
|
||||||
|
WITH PerQuadruple AS (
|
||||||
|
SELECT
|
||||||
|
pname.str AS pname,
|
||||||
|
IFNULL(ver.str, char(0x274C)) AS version,
|
||||||
|
plat.str AS platform,
|
||||||
|
IFNULL(ct.str, char(0x274C)) AS tag,
|
||||||
|
COUNT(DISTINCT sha256.hash) AS uq_sha256s
|
||||||
|
FROM
|
||||||
|
(
|
||||||
|
CudaArtifact AS ca,
|
||||||
|
Str AS pname,
|
||||||
|
Str AS plat,
|
||||||
|
Hash AS sha256 ON ca.pname = pname.id
|
||||||
|
AND ca.platform = plat.id
|
||||||
|
AND sha256.id = ca.sha256
|
||||||
|
)
|
||||||
|
LEFT JOIN Str AS ver ON ver.id = ca.version
|
||||||
|
LEFT JOIN Str AS ct ON ca.compat_tag = ct.id
|
||||||
|
GROUP BY
|
||||||
|
pname.id,
|
||||||
|
version,
|
||||||
|
platform,
|
||||||
|
tag
|
||||||
|
ORDER BY
|
||||||
|
uq_sha256s DESC,
|
||||||
|
pname.str,
|
||||||
|
version,
|
||||||
|
plat.str,
|
||||||
|
tag
|
||||||
|
)
|
||||||
|
SELECT
|
||||||
|
pname,
|
||||||
|
(
|
||||||
|
SELECT
|
||||||
|
GROUP_CONCAT(x, char(10))
|
||||||
|
FROM
|
||||||
|
(
|
||||||
|
SELECT
|
||||||
|
DISTINCT version AS x
|
||||||
|
FROM
|
||||||
|
PerQuadruple
|
||||||
|
WHERE
|
||||||
|
pname = pq.pname
|
||||||
|
ORDER BY
|
||||||
|
x
|
||||||
|
)
|
||||||
|
) AS versions,
|
||||||
|
(
|
||||||
|
SELECT
|
||||||
|
GROUP_CONCAT(x, char(10))
|
||||||
|
FROM
|
||||||
|
(
|
||||||
|
SELECT
|
||||||
|
DISTINCT platform AS x
|
||||||
|
FROM
|
||||||
|
PerQuadruple
|
||||||
|
WHERE
|
||||||
|
pname = pq.pname
|
||||||
|
ORDER BY
|
||||||
|
x
|
||||||
|
)
|
||||||
|
) AS platforms,
|
||||||
|
(
|
||||||
|
SELECT
|
||||||
|
GROUP_CONCAT(x, char(10))
|
||||||
|
FROM
|
||||||
|
(
|
||||||
|
SELECT
|
||||||
|
DISTINCT tag AS x
|
||||||
|
FROM
|
||||||
|
PerQuadruple
|
||||||
|
WHERE
|
||||||
|
pname = pq.pname
|
||||||
|
ORDER BY
|
||||||
|
x
|
||||||
|
)
|
||||||
|
) AS tags,
|
||||||
|
(
|
||||||
|
SELECT
|
||||||
|
SUM(uq_sha256s)
|
||||||
|
FROM
|
||||||
|
PerQuadruple
|
||||||
|
WHERE
|
||||||
|
pname = pq.pname
|
||||||
|
) AS uq_sha256s
|
||||||
|
FROM
|
||||||
|
PerQuadruple AS pq
|
Loading…
Add table
Add a link
Reference in a new issue