mirror of
https://github.com/rocky-linux/peridot.git
synced 2024-11-18 11:21:25 +00:00
226c0f4c30
Signed-off-by: Mustafa Gezen <mustafa@ctrliq.com>
108 lines
3.6 KiB
Go
108 lines
3.6 KiB
Go
// Copyright (c) All respective contributors to the Peridot Project. All rights reserved.
|
|
// Copyright (c) 2021-2022 Rocky Enterprise Software Foundation, Inc. All rights reserved.
|
|
// Copyright (c) 2021-2022 Ctrl IQ, Inc. All rights reserved.
|
|
//
|
|
// Redistribution and use in source and binary forms, with or without
|
|
// modification, are permitted provided that the following conditions are met:
|
|
//
|
|
// 1. Redistributions of source code must retain the above copyright notice,
|
|
// this list of conditions and the following disclaimer.
|
|
//
|
|
// 2. Redistributions in binary form must reproduce the above copyright notice,
|
|
// this list of conditions and the following disclaimer in the documentation
|
|
// and/or other materials provided with the distribution.
|
|
//
|
|
// 3. Neither the name of the copyright holder nor the names of its contributors
|
|
// may be used to endorse or promote products derived from this software without
|
|
// specific prior written permission.
|
|
//
|
|
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
// AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
// IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
// ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
|
|
// LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
// CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
// SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
// INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
// CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
// ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
// POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
package obsidianimplv1
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"github.com/ory/hydra-client-go/client"
|
|
"github.com/sirupsen/logrus"
|
|
"google.golang.org/grpc"
|
|
"net/url"
|
|
commonpb "peridot.resf.org/common"
|
|
obsidiandb "peridot.resf.org/obsidian/db"
|
|
obsidianpb "peridot.resf.org/obsidian/pb"
|
|
"peridot.resf.org/servicecatalog"
|
|
"peridot.resf.org/utils"
|
|
)
|
|
|
|
type Server struct {
|
|
obsidianpb.UnimplementedObsidianServiceServer
|
|
|
|
log *logrus.Logger
|
|
db obsidiandb.Access
|
|
hydra *client.OryHydra
|
|
}
|
|
|
|
func NewServer(db obsidiandb.Access) (*Server, error) {
|
|
adminURL, err := url.Parse(servicecatalog.HydraAdmin())
|
|
if err != nil {
|
|
return nil, fmt.Errorf("could not parse hydra admin url, error: %s", err)
|
|
}
|
|
|
|
hydraSDK := client.NewHTTPClientWithConfig(nil, &client.TransportConfig{
|
|
Schemes: []string{adminURL.Scheme},
|
|
Host: adminURL.Host,
|
|
BasePath: adminURL.Path,
|
|
})
|
|
|
|
return &Server{
|
|
log: logrus.New(),
|
|
db: db,
|
|
hydra: hydraSDK,
|
|
}, nil
|
|
}
|
|
|
|
func (s *Server) interceptor(ctx context.Context, req interface{}, usi *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (interface{}, error) {
|
|
n := utils.EndInterceptor
|
|
|
|
return n(ctx, req, usi, handler)
|
|
}
|
|
|
|
func (s *Server) Run() {
|
|
res := utils.NewGRPCServer(
|
|
&utils.GRPCOptions{
|
|
Interceptor: s.interceptor,
|
|
},
|
|
func(r *utils.Register) {
|
|
endpoints := []utils.GrpcEndpointRegister{
|
|
commonpb.RegisterHealthCheckServiceHandlerFromEndpoint,
|
|
obsidianpb.RegisterObsidianServiceHandlerFromEndpoint,
|
|
}
|
|
|
|
for _, endpoint := range endpoints {
|
|
err := endpoint(r.Context, r.Mux, r.Endpoint, r.Options)
|
|
if err != nil {
|
|
s.log.Fatalf("could not register handler - %v", err)
|
|
}
|
|
}
|
|
},
|
|
func(r *utils.RegisterServer) {
|
|
commonpb.RegisterHealthCheckServiceServer(r.Server, &utils.HealthServer{})
|
|
|
|
obsidianpb.RegisterObsidianServiceServer(r.Server, s)
|
|
},
|
|
)
|
|
|
|
defer res.Cancel()
|
|
res.WaitGroup.Wait()
|
|
}
|