/* * Copyright (c) 2022-2024, NVIDIA CORPORATION. All rights reserved. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #pragma once #include "tensorrt_llm/executor/types.h" #include "tensorrt_llm/layers/baseLayer.h" #include "tensorrt_llm/layers/penaltyLayer.h" namespace tensorrt_llm::layers { template class DynamicDecodeLayer : public BaseLayer { using Base = BaseLayer; public: DynamicDecodeLayer(executor::DecodingMode const& mode, DecoderDomain const& decodingDomain, std::shared_ptr bufferManager); void setup(runtime::SizeType32 batchSize, runtime::SizeType32 beamWidth, runtime::ITensor::SharedConstPtr batchSlots, std::shared_ptr const& setupParams, std::shared_ptr const& workspace) override; void forwardAsync(std::shared_ptr const& outputs, std::shared_ptr const& inputs, std::shared_ptr const& workspace) override; void forwardSync(std::shared_ptr const& outputs, std::shared_ptr const& inputs, std::shared_ptr const& workspace) override; //! @returns workspace needed for this layer in bytes [[nodiscard]] size_t getWorkspaceSize() const noexcept override; void disableLookahead(DecoderDomain const& decoderDomain, SizeType32 batchSize, TensorConstPtr batchSlots, std::shared_ptr const& baseSetupParams, std::shared_ptr const& workspace); private: void allocateBuffer(); void initialize(); void initializeLayers(); void prepareIdsPtrs(std::shared_ptr const& outputs, BufferConstPtr batchSlots, runtime::SizeType32 batchSize, runtime::SizeType32 beamWidth, runtime::SizeType32 maxSeqLen); void prepareOutputData(std::shared_ptr const& outputs, BufferConstPtr batchSlots, runtime::SizeType32 batchSize, runtime::SizeType32 maxBatchSize, runtime::SizeType32 beamWidth, runtime::SizeType32 maxSeqLen, runtime::SizeType32 maxTokensPerStep, bool outputLogProbs, cudaStream_t stream); private: using Base::mDecoderDomain; std::vector> mLayers; executor::DecodingMode mDecodingMode; TensorPtr mZeroParentIdsDevice; TensorPtr mOutputIdsPtrHost; TensorPtr mParentIdsPtrHost; TensorPtr mOutputIdsPtrDevice; TensorPtr mParentIdsPtrDevice; bool mHasDiffRuntimeArgs{false}; bool mOutputLogProbs{false}; runtime::SizeType32 mCyclicStep{0}; runtime::SizeType32 mRuntimeMaxSeqLen{0}; runtime::SizeType32 mConfiguredBeamWidth{-1}; }; } // namespace tensorrt_llm::layers