-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[WIP] feat: PDP #227
base: feat/market
Are you sure you want to change the base?
[WIP] feat: PDP #227
Conversation
07a5b22
to
5c3f546
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I've mostly put in comments to check for understanding before I start writing code. But generally LGTM -- looks like it's compliant with the API we agreed on :)
) | ||
|
||
// verifyJWTToken extracts and verifies the JWT token from the request and returns the serviceID. | ||
func (p *PDPService) verifyJWTToken(r *http.Request) (string, error) { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@magik6k if I am reading this right, I send a JWT with a claim of "service" = "storacha" and then I need to make sure the JWT is signed by a private key that corresponds to a public key that was previously inserted into the pdp_services
table with probably the RPC api below to add PDP Service (i.e the SP called "AddPDPService" locally using their running node -- assuming "AddPDPService" is not available on the public endpoint)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yeah, see
Line 201 in 54af620
func createJWTToken(serviceName string, privateKey *ecdsa.PrivateKey) (string, error) { |
cuhttp/server.go
Outdated
@@ -204,5 +215,10 @@ func attachRouters(ctx context.Context, r *chi.Mux, d *deps.Deps) (*chi.Mux, err | |||
} | |||
ipni_provider.Routes(r, ipp) | |||
|
|||
if sd.EthSender != nil { | |||
pdsvc := pdp.NewPDPService(d.DB, d.LocalStore, must.One(d.EthClient.Get()), sd.EthSender) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
this is essentially what's adding all the PDP routes correct (i.e. /proof-sets
, POST to /piece
)
// Collect pieceInfos for subroots | ||
pieceInfos := make([]abi.PieceInfo, len(addRootReq.Subroots)) | ||
|
||
var totalOffset uint64 = 0 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
am I correct that reading this code right here provides indirect instructions on how to aggregate all the subroot pieces into a root CID? IOW, I can assume that I can basically want to write aggregation code that will satisfy this code by generating the same RootCID? (which appears to be all the pieces concatenated together).
Also it would appear non-ffi is a fork of go-commp-utils -- is this public?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, the subroot aggregation is the same as in filecoin pieces (CommP -> CommD)
The non-ffi thing is the same code as the Rust logic, reimplemented by Riba in Go.
There is one added constraint in PDP - pieces must be sorted by size in descending order - this way no inter-piece padding is needed
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
ooh good call out good to know!
pdp/handlers_upload.go
Outdated
|
||
// Parse request body | ||
var req struct { | ||
PieceCID string `json:"pieceCid"` |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It looks like you're not requiring piececid v2 here, not in the put handler. Long term, I think we should require this. From the standpoint of the SP, I imagine them to want to filter at the piece level, and I imagine one of the main filter critieria to be size. They should know exactly what size piece they're about to accept before they accept it.
Unless I'm wrong, a PieceCID v1 doesn't specify that, and absent a size parameter, they could be accepting anything from 32 bytes to 256 MB.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, PieceCIDv2 is still todo, can quickly update if that's a blocker
"github.com/filecoin-project/curio/lib/storiface" | ||
) | ||
|
||
type PieceParkReader struct { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is what finds and reads PDP pieces for get requests correct?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yeah, PiecePark is the general "raw piece" storage in Curio, we eventually intend to serve all retrievals from it (instead of from the fr32-padded "unsealed" sector files), but PDP is the first place which will actually do reads directly from it.
am I correct that the retrieval handler already has been modified to support PDP pieces through the parked piece reader? |
am I correct in believing that the next step on the API is that add root will send a location header to a creation status API for the root, then that will be used to get the root ID which is then possible to pass to the root detail and delete root APIs? Not crazy urgent just want to know. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Github thinks I have an unsubmitted review here and can't discard it..
Yes, correct, retrievals work already. No ACL mechanism yet, but it's not going to be hard to implement once we have a spec (you can see that all retrieval requests quite quickly become aware what piece they are reading from + ACLs should be quite cachable) |
Other than delete all the things should be there. Operations on PDP proofsets (other than create) you should probably watch through eth events. I can also make the message-sending endpoints output the message hash somehow, but you still want to watch for all the events to not miss any contract changes made by the SP |
fa8d937
to
a3a1c3a
Compare
Upon further review, it's possible we have a design issue here that we need a solution for. Please let me know if I am understanding the sequence of events correctly when a user uploads a pdp piece:
Why this matters: a basic gaurantee of all hot storage, which storacha tries to offer, is read-on-write. The basic expectation is from the users point of view is that when an upload finishes, it should be retrievable. One thing that is not clear I'd like to clarify: are steps 3 & 4 likely to run in a couple seconds or much longer? If it's only a few seconds, we can work around this. Fortunately our upload process includes a confirmation step AFTER the upload, and what we can do is essentially block the confirmation until step 4 completes. But that only works if it's a short period of time before the piece is fully retrievable. If not, we really need to revisit the architecture and find a way to shorten the process. |
Follow-up: a pull based approach for checking for piece cid -- #304 |
* feat(pdp): add handler to check for piece cid status * feat(pdp): add handler to check for upload exists
Allows accepting JWTs that are signed with Ed25519 as well as ECDSA
x509 does not parse to pointer type on ed25519
* Update PDP to new proof scheduling WIP * add listener addr to pdp tables * generate bindings for new proving_schedule iface * Complete draft of new challenge window scheduling * Try to get tests building * No fancy ALTER tables * Review response * Fix * Fix sql syntax --------- Co-authored-by: zenground0 <[email protected]>
Reading:
Roadmap: