Skip to content

Commit

Permalink
fix the head file
Browse files Browse the repository at this point in the history
  • Loading branch information
lambda7xx committed Feb 9, 2024
1 parent 0cdfb1a commit a7ef500
Show file tree
Hide file tree
Showing 5 changed files with 17 additions and 14 deletions.
2 changes: 1 addition & 1 deletion deps/legion
Submodule legion updated from 15b23c to 626b55
2 changes: 1 addition & 1 deletion deps/nccl
Submodule nccl updated 142 files
12 changes: 12 additions & 0 deletions include/flexflow/ops/fused.h
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,18 @@

namespace FlexFlow {

// declare Legion names
using Legion::Context;
using Legion::coord_t;
using Legion::Domain;
using Legion::Future;
using Legion::LogicalPartition;
using Legion::LogicalRegion;
using Legion::Memory;
using Legion::PhysicalRegion;
using Legion::Runtime;
using Legion::Task;

class FusedOp;
class FusedOpMeta {
public:
Expand Down
1 change: 1 addition & 0 deletions llama.sh
Original file line number Diff line number Diff line change
@@ -0,0 +1 @@
./build/inference/incr_decoding/incr_decoding --ll:util 8 -ll:cpu 4 -ll:fsize 20000 -ll:zsize 30000 -llm-model meta-llama/Llama-2-7b-hf -cache-folder /home/xiaoxias/weights --fusion -ll:gpu 4 -prompt /home/xiaoxias/cudagraph/inference/prompt/chatgpt.json -tensor-parallelism-degree 4 > llama2.log 2>&1
14 changes: 2 additions & 12 deletions src/ops/fused.cu
Original file line number Diff line number Diff line change
Expand Up @@ -47,17 +47,6 @@
#include "flexflow/ffconst_utils.h"

namespace FlexFlow {
// declare Legion names
using Legion::Context;
using Legion::coord_t;
using Legion::Domain;
using Legion::Future;
using Legion::LogicalPartition;
using Legion::LogicalRegion;
using Legion::Memory;
using Legion::PhysicalRegion;
using Legion::Runtime;
using Legion::Task;

OpMeta *FusedOp::init_task(Task const *task,
std::vector<PhysicalRegion> const &regions,
Expand Down Expand Up @@ -1127,7 +1116,7 @@ __host__ void FusedOp::capture_graph(Task const *task,
output_accessors_to_save.push_back(output_accessor[i + ooff]);
}
assert(task->index_point.get_dim() == 1);
int shard_id = task->index_point.point_data[0];
shard_id = task->index_point.point_data[0];
FusedOp::save_inference_tensors_to_file(metas->meta[op],
shard_id,
bc,
Expand Down Expand Up @@ -1176,6 +1165,7 @@ __host__ void
bc->num_active_tokens(),
bc->num_generation_tokens > 0);
int scenario = 0;
int shard_id = task->index_point.point_data[0];
auto it = metas->graph_collections.find(graph_params);
if(it != metas->graph_collections.end()) {
instance = it->second;
Expand Down

0 comments on commit a7ef500

Please sign in to comment.